0 ratings0% found this document useful (0 votes) 517 views514 pagesDP 203
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content,
claim it here.
Available Formats
Download as PDF or read online on Scribd
=XAMTOPICS
- Expert Verified, Online, Free.
‘ custom View Setings
Topic 1 - Question Set 1You ave table in an Azure Synapse Analytics dedicated SQL poo, The table was created by sing the following Transact-SQL statement,
CREATE TABLE [dbo] . [DimEmployee] (
[SmployeeKey] [int] IDENTITY (1,1) NOT NULL,
(EmployeeTD] [int] NOT NULL,
[FirstName] [varchar] (100) NOT NULL,
[LastName] [varchar] (100) NOT NULL,
(JobTitle] [varchar] (100) NULL,
[LastHireDate] [date] NULL,
(Streetaddress] [varchar] (500) NOT NULL,
[City] [varchar] (200) NOP NULL,
[StateProvince] [varchar] (50) NOT NULL,
(Portalcode] [varchar] (10) NOT NULL
)
You ned to alter the table to meet he following requiemens:
> Ensure that users can identify the current manager of employees.
c» Support creating an employee reporting hierarchy for your entre company.
-» Provide fast lookup ofthe managers’ attributes such as name and job tte
Which column should you add tothe table?
{A [ManagerEmployee] [small] NULL
8. [ManagerEmployeekey|[smalint] NUL
6. [Manager mployeekey] int] NULL
. [Managerttame][vrchar(200) NULL
Correct Answer: ©
‘We need an extra column to identity the Manager. Use the data type asthe EmployeeKey column, an int column,
Reference:
hnepsffdoes microsoft com/en-us/analysisservces/tabular-models/hierarchies-ssastabular
Community vote dtrbuton
© i Jakibick (aT ITS) 10 months 90
swerC, Smallnt eliminates A and . But would name the feld [ManagerEmployeelD] int] NULL since it should reference EmployeelD, not
Employeekey since this one i IDENTITY.
pote 14 tines
DB Chandra2099 5 wecks, 2 days ago
IDENTITY keyword suggests that its a surrogate key and it's 2 good practice to have a separate field forthe surrogate key which isnot 2
business fil like emp no or emp name
puted Times
© & anures0dtest 5 months 999
Why is smalint not possible. would sill support more than enough range to list all the employees inthe company and save space + speed up
the lookup,
puted 7 mes
© & anures0otest 5 months 290
Salt i upto 32,767: he
vers
‘pote I nee
© & alexleonvatencia TTT) 11 months, 3 weeks 299
Lieespuesta es correcta
puted 11 tines
© & Mansoorsen9s (T=) day, shou a
Selected Answer €
s/)docsimicrosott.com/en-us/s/t-sq/data-types/nt-bigin-smalint-and-tinyit-transact-sviepte T ines
© amipanah 3 weeks, 24ays 299
Guys i passed the exam today and i would say more than 90% ofthe questions was exactly the same as this! Good job Examtopic and community.
upted tines
2 & hobeyor 2 months, 2 weeks 990
iknkis
pte Times
© A examtopieseap 3 months, 3 weeks ago
correct,
pte Times
1D & erazybobs 3 months, 3 weeks ago
Soundsright
pte Times
© A Hansen 5 montis go
Answer Cis
pote Times
© & Nikes015 montis ago
“Answer. Looking atthe table itis bul for OLAP where Employeekey wll be serrogate key so basing on thatthe ManagerEmployeeKey Looks
4204, fype shoule match and NULL constraint shoul be apalied (big boss hae ne Bose)
‘puted 2ines
© Mem $ months, 1 week ago
correct ans iC
pte Times
© & mithomas 6 montis a90
€: ManagerEmployeeKey INT NULL
Remember, his should be NUL since there are employees that may not have a manager Ge, the boss).
This design is usually seen with recursive CTE proof of concep,
pte Tver
© & Dotty 6 months, 3 weeks ago
Selected Answer C
pte Times
BA Egocentie 7 months, 2 weeks ago
Cis te correct answer
‘puted times
© & boggy011 7 months, 3 weeks ago
Selected Answer €
puted Times
BW temace & months, 1 week ago
Comect answers
‘puted T nes
© & Guincimund & mons, 2 weeks ago
answer's
‘puted T ines
© & NeerajKumar 8 months ago
Comect Ans iC
puted 2 inesYou have an Azute Synapse workspace named MyWorkspace thal contain an Apache Spark database named mytestdb
You run the following command in an Azute Synapse Analytics Spark pool in MyWorkspace
‘CREATE TABLE mytestdb.myParquetTabl
ployeelD in,
EmployeeName sting,
EmployeeStartDate date)
USING Parquet
You then use Spark to insert row into mytestéb myParquetTable.Therow contains the folowing data
EmployeeName | EmployeeID | EmployeeStartDate
Alice 24 2020-01-25
(ne minte later you execute the following query rom a serverless SQL poolin MyWorkspace
SELECT EmployeeiO-
FROM mytestb. dbo myParquetTable
WHERE EmployeeName = Alice’;
What will be returned bythe query?
Am
B.anerror
null value
Correct Answer: A
‘Once a database has been created by a Spatk jo, you can create tables in t with Spark that use Parquet as the storage format. Table names
will be converted to lower case and need tobe queried using th lowercase name, These tables wll mediately become availabe for querying
by any of the Azure Synapse workspace Spark pools. They can also be used from any ofthe Spark jobs subject to permissions
Note: For extemal tables, since they are synchronized to serveress SQL pool asynchronously there will be a delay util they appear.
Reference:
upsifdocs.mierosoft.com/en.us/azure/synapse-analytics/metadata/table
Community vote dstrbution
© & deatrypo [in weedd) 1 mort, 1 week ago
‘Answer i 8, but not because of the lowercase, The case has nothing to do withthe error.
Ifyou look attentively, you wll notice that we create table mytestdb myParquetTable, but the select statement contains the reference to table
mytestdb doo myParquetTable(!!- dbo)
Here isthe eror message | got
Error spark catalog requires a single-part namespace, but got [mytestdb, dbo]
‘puted tines
© & gerie1979 (aS) month ago
Ici test waited for one minute and tied the query ina serverless sal pool and received 24 asthe result, 50 I don't understand that B has been
voted so much because the answer i A) 24 without a doubt
pote tines
© & maximilianogarcias 2 weeks, 6 ays 90
Did you tried the same query that is presented here? with “mytesta.clbo.myParquetTse"??
puted 3 mas
© & Billybobo604 [Terie days, 21 nous a90
‘there's another consideration, as to know the way the record is inserted. when this is dane specifying the columns it will succeed but witha select *
itwon' hence the € would be the answer, But let's assume the racer i inserted correct.
puted T nes
© gabriethuka 1 weak, 1 day ag0
| vie this in Azure Synapse an it returns an error Removing the "dbo" part works So answer i 8.
pote Times
1 & abrar8H0 4 days, sous agoDid you mentioned table name in Lowercase inthe query?
‘ted T nes
BA aad 1 week 4 days 299
dash is just a typo, simply ignore it
puted Times
© & withrocky 2 weeks, 2 days ago
“abo in the select statement wil sultan error. If this SQL statement i given to check presence of mind, then answer is B orf the dbo it
Statement i 9 typo then answer is A
‘pted2trmes
sal
© A ashwin Joshias 4 wesks, 1 day ago
‘Answer is shown a8 8 here. 'ied to recreate the scenario and when we execute the select query t's throwing an errar because of hyphen symbol
only but not because of upper & lowercase
So lam not sure f hyphen is typo error. yes then answer should be A else
Suppasein select query if dbo object mising then error will occur
‘pted 7 ines
© & ted0809 1 mont ago
We are not asking ower or upper case here
Look atthe bigger picture,
Returning 28
pote Times
© igormmpinto 1 month, 2 weeks ago
Answer is
But think thats because in CREATE TABLE statement we have “myjtestdlsmyParquetTablo" and in SELECT statement we have
mmytestdb doo myParquetTable
puted? mes
© Winwan33 2 montis apo
Answer is 8
it will return an errr. table name wil be converted to lowercase
Iitpsd earn microsoft comfen-us/szure/synapse-anaytis/metadata/able
pve tnes
© Jawidkaderi 2 mons 990
‘The query is correc, The spelling is correct. | am not sure why the upper an ower casei of any concern her; it has to do with the collation S, if
ignore the dash aftr the end of SELECT line the answer would be 24
pte 2 tines
© & Man01997 2 months ago
itiethe answer
puted T tines
BM anks84 2 months, 3 weeks 290
Once you create has been created, the table names willbe converted to lower case and need to be queried using the lowercase name,
Hence, the answer is B (Fron,
puted inas
2M anks84 2 months, 3 weeks 290
‘Once the table has been created
pote Ties
Ignore the Type above
© & debarun 2 months, 2 weeks 290
Error. Table names wll be converted to lower case and need to be queried using the lower case name
‘pte T nas
© & debarun 2 months, 2 weeks 290
Eror. Table names willbe converted to lower case and need to be queried using the lower case name
pte Times
© A Remedios79 3 mortis 290
since you wrote the table name in mixed awer and uppercase it should give you error because it expects lowercase table name
FROM mytestdi.dbo myParquetTable
pte tines© & rafeek00007 2 months ago
asters B.A eor will occur
pte TimesRAG DROP.
You have a table named SalesFact in an enterprise data warehouse in Azure Synapse Analyt. SalesFact contains sales data fromthe past 36
‘months and has the following characteristics:
= Is partitioned by month
“» Contains one bilion rows
“> Has clustered columnstore index
[tthe beginning of each month, you need to remove data ftom SalesFact that Is older than 36 months as quickly as possible,
Which three actions should you perform in sequence ina stored procedure? To answer, move the appropriate actions from the Ist of actions to the
answer tea and arrange them inthe correct order.
Select and Pace:
Actions
Switch the partition containing the stale data
from SalesFact to SalesFact_ Work.
Truncate the partition containing the stale data
Drop the SalesFact_ Work table.
Create an empty table named SalesFact Work
that has the same schema as SalesFact.
Execute a DELETE statement where the value in
the Date column is more than 36 months ago.
Copy the data to a new table by using CREATE
TABLE AS SELECT (CTAS),
Correct Answer:
Actions
Answer Area
Answer Area
‘Switch the partition containing the stale data
from SalesFact to SalesFact_ Work.
Create an empty table named SalesFact Work
that has the same schema as SalesFact.
Truncate the partition containing the stale data.
‘Switch the partition containing the stale data
from SalesFact to SalesFact Work.
Drop the Salestact Work table.
Drop the SalesFact Work table.
Create an empty table named SalesFact Work
that has the same schema as SalesFact.
Execute a DELETE statement where the value in
the Date column is more than 36 months ago.
Copy the data to a new table by using CREATE
TABLE AS SELECT (CTAS).
‘Step 1: Create an empty table named SalesFact.
ork that has the same schema as SalesFac.
Step 2: Switch the partition containing the stale data from SalesFact to SalesFact Mork.
SQL. Data Warehouse supports partition spliting, merging, and switching. To switch partitions between two tables, you must ensure that the
patitions algn on thir spective Boundaries and that the tale defitions match
Loading data into partitions wih patton switching isa convenient way stage new data ina table that snot visible to users the switch nthe
new data
Step 3: Drop the SalesFact_ Work table
Reference:
hitps:ifdoes microsoft com/enus/azure/sal-data.warehouse/sqldats-warehousetabes-patition
© & seta (RTOS) year, 2th 99Given answer D A Cis coreect
pte 33 tines
BB sik year 2monits ago
Yes. Once the partion
poted2times
1 & Jawidkaderi ox)? months ap0
very interesting questions
Every partion has a name, which indicated by the mmY¥YY perhaps.
So, if we know the name ofthe partion, we can drop that parction dirty:
DROP PARTITION SCHEME parton scheme name [>
However, if there i an index onthe table DOPR Patton will not work So, the iis correct.
Dac.
‘puted T ines
1 with an empty partition itis equivalent to truncating the partition from the original table
© & pmeos 3 months, 1 weck ago
Answer is F-A-
https//docs miroso
posted times
comes-es/achive/blogs/apsblag/azuresel-dw-performance-ctasparttion-switching-vs-updatedelete
© & supriyako 1 month, 1 week 290
F seems wrong ast says CTAS to copy the data
ted Tee
© & pme08 3 moni, 1 week go
Dis incorrect because we also need to copy the data onto the new table
voted tines
© & Deekshat234 3 months, 2 weeks ago
correct ans
‘puted T nes
2A mithoma3 6 months ago
DAC
‘Azure Synapse does not support truncating partitions. Currently, that
pote Times
feature is only tied to MS SOL Server.
© & Dathy 6 months, 2 weeks 290
Stop 1: Create an empty table named SalesFact work that has the same schema as SalesFact.
Step 2: Switch the partion containing the stale data from SalsFact to SalesFact Work
Step 3: Drop the SalesFact Work able
puted Times
2B dudewit7 months, 3 weeks 250
DAC isthe right option
For more information this doc discusses exactly this example: https/does microsoft com/en-us/azure/synapse-analytics/sql dat
data-warehouse-tablesparttion
pte Times
© & theezins months ago
Why net included deb
puted Times
9 sales data older than 36 months whichis mentioned in question?
© & RamGhase 9 months, 3 weeks 290
i could not understand how answer handled to remave data before 36 month
puted T nas
rard 9 months, 3 weeks 290
you have to move the partitions that contains the date before 36 months
voted 3 nas
| 0 onthe, 7 week 290
DACiscom
pote tines
© indomanish 11 mons ago
Patton switching help us in loading large dataset quickly. Nat sure iit will help in purging data ae well
puted 2imes
© & Sabadamal2010atGmall 11 montis, 1 week 99
Given answers correct
‘puted ines© & covllmalls yea, 1 month ag0
DAC is ort
pte nee
© A uit 1 yea: s months ag0
truncate partitions even quicker, why isnt that the answer, ifthe data is dropped anyway?
‘puted 3 mes
© & BlackMal year 3 months ago
This think it should be the answer
ted nse
© & yolapa1172 1 year, 1 month ago
There i no way to truncate partitions in Synapse. Partitions don’ even have names and you can't reference them by valve,
‘pote tines
© & poomipy 1 yea 4 months ape
what is
‘poted times
1 correct answer for this?
© & AnonzuredataEngineer + yea, 4 months 290
Seems like it should be:
pte T nes
1 & dragos.dragos62000 1 year Smontis 290
Coneet
ted | neeYou have files and folders in Azure Data Lake Storage Gen? for an Azure Synapse workspace as shawn inthe following exhibit
‘Mtoptolder/
| Filer.csv Htoldert/ older2/_| Filed csv
I |
File2.csv Fileacsv |
You create an exteral table named ExtTable that has LOGATION= /topflder
When you query ExtTableby using an Azure Synapse Analytics serverless SQL pool, which ile are returned?
‘A. Filed and File8.esv only
8. Fle.csv and Flees only
Fle. cs File? Fle, and Fled.esv
Filet .esv only
Correct Answer: C
Toruna TSQL query over set of les within a folder or st of folders while treating them as a single entity or rows, provide a path oa folder
orapattem
(using wildcards) over @ set of files or folders.
Reference:
hntpsffdacs microsoft com/en-us/azute/synapseanalytes/sql/query-datestorage¥query-multiple-les-orfolders
Community vote dstrbution
© & Chilem1900 ama) 1 yea, 7 months ago
believe the answer should be 8,
In case ofa serveress pool a wildcard should be added to the location.
https://decs microsoft com/en-us/azure/synapse-analytics/sq/davelop-tables-exteral-ables?tabs=hadoop targuments-ceate-external-table
puted 91 times
© & captainpike 1 yas, 1 month 90
| tested and proove you right, the answer is 8. Remind the question is refering to serverless SOL and not dedicated SQL pool. “Unlike Hadoop
fevteral tables, native external tables don't return subfolders unless you specify atthe end of path In this example, if
LOCATION="/websdata/, a servriess SQL pool query, ill retun cows from mydata at. It won't return mydata? it and mydata3ut because
theyre located in a subfolder Hadoop tables wil return ll files within any subfolder”
voted 20 tines
© & alana GTS 1 year, months ago
“Servrless SQL pool can recursively traverse folders only if you specty/* atthe end ofp
hitps//decs microsoft com/en-us/azure/synapse-analyis/sql/query-folders-multiple-sefles
upeted 21 times
2 & Preben year Smonths ago
|when you are quoting fom Microsoft documentation, do not ADD in words tothe sentence Only"is not used
ted 12 times
© & captaipike 1 year, 1 month go
‘The answer is B however. Icould not make
pote tines
© & brohanyu (mea) 1 day 1Shous 399
Bis correct
‘pte T nas
ta work somebody?
© rohitbinman 1 woe, 4 days 290
“native external tables don’t return subfolders unless you specify /* atthe end of path In this example, f LOCATION=""webdata/’, a serverless SOL
pool query, wil return routs fram mya xt. It ont return mycata2. and mycsta8 tt because they're located in a subfolder, Hadoop tables will
Fetuanal files within any sub-folder
https://learn microsof.com/en-us/saure/synapse-analtics/sql/develop-tables-exernal-tablesabs=hadooparguments-create-exteral-tablepte T ines
© gerret979 1 monn, 1 week 290
tested the solution out in Azure Synapse with ll the necessary steps connected tothe Data Lake Gen2 storage account and answer B is correct
the subdirectories are not taker into account
pote Times
DB dawidkaderi 2 months 290
kind of agree with captanpike, However, | have my reason ase:
lam reading the questions as such
the table created inside te /topfolder/
hich means the new table isin the same linea: Flees, folderl/ flder2, Filedesv
So, think A] isthe correct answer. FleZesv and File3.csv
‘puted 7 nes
© & smsme323 2 months, 1 week ago
ned wildcard to acces al fie,
‘puted T ines
2 gerret979 monn ago
correct, only with location
‘puted T ves
‘opfolder/™ it ill hve access to all the subfolders
2M anks84 2 months, 3 wecks 290
Given answer C is Correct
‘puted Times
© & yyphinh 2 months, 1 week ago
Tagree with @temace, C
‘pted 2 ines
© & examtopieseap 3 months, 3 weks ago
correct 8
‘puted? ines
2 & Razat2 «months ago
Unlike Hadoop externa tables, native externa tables don’ return subfolders unless you specify /* at the end of path. I this example, if
LOCATION="/webdata, a serverless SQL pool query ill return rows from mydata txt. I won etum mydata2 tx and mydata8 txt because they're
located in subfolder. Hadoop tables wll return al les within any sub-older,
upvted 3 tines
DB danieys «months, 1 week ago
Bis the correct Same example in: ttps:/locs microsoft camyen-us/sqlt-sqU/statements/create-estemaltabl
Intestlipreservesview=truesitabe-serverest
pote Times
sact-svlew=szure-sald
BA Razat2 4 months, 2 weeks 290
(Seems from tis document that Answer is ’C"jln SQL Server, the CREATE EXTERNAL TABLE statement creates the path and folderft doesn't
Slready exist. You can then use INSERT INTO ta export data from local SQL Server table o the extemal datasource. For more information, £66
PolyBase Queries
you specty LOCATION to be 8 folder,» PolyBase query that selects from the exteral table wil retrieve files from the folder and lofts
subfolders. Just lke Hadoop, PolBase doesnt return hen folders. also doesn return files for which the fie name begins wth an undetline C)
ora period ()
In this example if LOCATION="/webdata, a PolyBase query wil return rows from mydata kt and mydat2.xt.K wont return mydatad it because
its fein iden folder. And it won return hidden tat because #3 hidden fle
Recursive data for extemal tables
htosflocs mirosofe com/en-us/sqlt-sql/statements/creste-extern-tabls
‘puted Times
© Nishikag 5 months, 1 week 200
‘Answer Bn serveress SQL pools must be specified /* atthe end of the location path, In Dedicated pool the folders are alwasy scanned
recursively,
https//decs microsoft com/en-us/azure/synapse-analytis/sq/dvelop-tables-exteral-tables7abs=hadoop
pted times
ransact-sol'view=sal-server-verlS&tabs=dedicated
BA temace 5 months, 1 week 3g0hutps//docs microsoft com/en-us/sqlt-sqistatements/create-extema-table-transac-sqlview=sql-server-verISB&tabs=dedicatedtlocation--
Folder-or flepath
pote 5 ines
BB amira404 sons, 1 neck 90
Cis comect as mentionned in the official documentation which showcase a similar example :tps//does mierosoftcony/en-us/sqVt=
sal statements/erete-external-table-transact-sqlview=sql-serverover Sétabs-decietedvlocation--flder_orflepath
puted tines
B & Backy 6 months, 3 weeks e90
‘The question does not show the actual query so this sa problem
pte TimesHoTsPor
You ate planning the deployment of Azure Data Lake Storage Gen2.
You have the following two reports that will access the data lake:
ce» Report: Reads thre columns from ale that contains $0 column,
“+ Report: Queries a single record based on a timestamp.
You need to recommend in which format to store the data inthe data lake to support the reports, The solution must minimize red times,
\What should you recommend fot exch report? To answer, select the appropriate options inthe answer ata,
NOTE: Each correct selections worth one point
Hot Aes:
Answer Area
Report: v
| Avro
csv
Parquet
TV
v
Avro
csv
Parquet
TSV
Answer Area
Report: v
Avro
csv
Parquet
Correct Answer: TSV
Report2: v
Avto
csv
Parquet
TSV
Report: o8V-
CSV; The destination writes records as delimited data,
Report2: AVRO
‘AVRO supports timestamps.
Not Parquet, TSV: Not options for Azure Data Lake Storage Gen2,
Reference:
hitps/streamsets.com/dacumentaton/datacollector/latesthelp/datacollector/UserGulde/Destinations/ADLS-G2-D.html
© & alana GTI yea, 6 months ago
1: Parquet -colummorinted binary fie format
2: AVRO - Row based format, and has logical type timestamp
hitpsyfveutu be/trWthxeT3UY
pote 137 times
2 A nteox 1 week. 9 doys go‘Thanks forthe video share, this relly helps, Cheers,
‘voted T nes
© @ terajuana* yew: 5 months ag0
the web is fl of eld information. timestamp support has been added to parquet
‘voted 5 nas
1D & vadege 1 yea: 5 months ago
‘Ok butin 1st case we need only 3 of $0 columns. Parquet i columnar format In 2nd Avro because ideal for red fll ow
ited IStines
© anurestudent1498 7 months, 2 weeks ago
thisis caret.
ited nes
© & infor (HEaTIaEES) 1 yer, months 90
Shouldn't the answer fr Report 1 be Parquet? Because Parquet format is Columnar and should be best for reading a few columns only,
puted 71 tines
© & Deekahat 234 [eai==515) 2 months, 2 weeks ago
1. Parquet
2. AVRO
pted tines
© & ROLLINGROCKS 5 montis ego
Parquet - Avro makes more sense given the defintios given by Microsoft
pted2simes
© & temace Smosh, 7 week ape
1. Parquet 3 of 50 columns
2. Parquet. In case with single row work faster, than AVRO,
‘pote Times
BM main616 6 months, 3 weaks ago
1. sv (arson swjson support query accelerate by select specified rows hitps//docs microsoft com/en-us/azure/storage/blobs/data-lake
storage-query-accelerationtoverview
2awo
‘pte nee
© & Doth 6 months, 2 weeks ag0
‘Parquet
2: AVRO
‘puted 3 mes
© & Ralphtiang 8 months ago
Consider Parquet and ORC fle formats when the /O pattems are more read heawy or when the query patterns are focused on a subset of columns
inthe records
the Avro format works well with 9 message bus sich as Event Hub or Kafka that write multiple events/messages in succession
poted 3 times
© A ragp.87 10 months ago
1. Parquet
2.Awro
hitps//docs microsoft com/en-us/azure/storage/blobs/datalake-storage-best-pratices
"Consider using the Avo file format in eases where your VO patterns are mare write heavy, or the query pattems favor retrieving multiple rows of
recordsin teir entirety.
Consider Parquet and ORC file formats wien the V/O pattems afe more read heavy or when the query patterns are focused on a subset of columns
in the records
pte 8 tines
BM SebK 8 months, 2 weeks apo
‘Thank you
ated Ines
© & Mohammadkhubeb 10 months, 1 week 90
Why NOT csv in report ?
puted 2simes
2 & dauqueto 3 days, 18 hous ago
(CSVis not optimized for column reads
‘ote T nas
© & Sandipau 10 months, 3 weeks 290‘This has to be parquet and AVRO, got the answer from Udemy
‘pte 5 nas
© Mahesh.mm 11 months, week 290
1. Parquet
2.AVRO
‘poted 3 times
© A marcint212 11 months, 3 weeks ag0
‘The goals: The sol
tion must minimize read times.
| made small test on Databrck pls Datalake,
‘The same file saved as Parquet and Avro
9 min of records
Parquet -150 MB
‘Avro ~TOOMB.
Reading Parquet always 10 times faster that Avro,
I checked
= forall data or small range of data with condition
“aller enly one columa
So wil select option
Parquet
parquet
pte 2 ines
2 & devaer 11 sorts ag0
how can be foster ead is same as number of ends?
voted | nes
© & Oazypoppe 1 year 200
Solution says parquet isnot supported for als gen 2 but it actually is: httos//docs microsoft. comven-us/azure/data-factory/fermat-parquet
puted 3 times
© & noranathae yea, * month 390
An interesting ane complete sricle that explain the dfferent uses between parquet/avra/esv and gives answers ta the question
https://mecium coryssense-tech/esv-vs-parquet-vs-avto-cnoosing-the-ight-tal-fr-the-ightjob-79c3/5691 a8
pve ines
BB elimey + yeas, 4 months ago
hitps/luminousmen com/postbig-data-fle-formats
‘pte T nas
BB etimey 1 yeas, 4 months ago
Report 1 definitely Parquet
pote TimesYou ate designing the folder structure for an Azure Data Lake Storage Gen? contane.
Users wll query data by using a variety of services including Azure Databrcks and Azute Synapse Analytics serverless SQL pools. The data will be
secured by subject area, Most queries wll include data from the current year or current month,
Which folder structure should you recommend to support fast queties and simplified folder security?
‘A HSubjectAvea) DataSource) (OD) (MM)/(YYYY)/(FleData)VYYY) {MM} {0D}. csv
8, (ODY{MM)AYYYY) (SubjectArea} (DataSource) (FleData) VY) (MM) {00}.c8¥
AYYYYN(MM)/{OD)/{SubjectArea) DataSource) (FleData)VYYY) {MM} {00}. csv
1. SubjectArea) (DataSource) (¥Y¥Y}/(MM)/{00)/ (leat) {YYVY) {4M} (00}.cs¥
Correct Answer: D
‘There's an important reason to put the date atthe end of the directory structure, I you want o lock down certain regions or subject matters to
users/groups then you can easily da so wih the POSIX permissions. Otherwise, if there was a need to restrict a certain security group to
viewing jst the UK data or certain planes, with the date structure in font a separate permission would be required for numerous directories
under every hour directory, Additionally having the date structure infront would exponentially increase the numberof directories as time went
Note: In oT workloads, there can be a great deal of data being landed inthe data stove that spans across numerous products, devices,
organizations, and customer. I's important to pe-plan the diectory layout for organization, secu, and efficient processing ofthe data for
doworstream consumers. A general template to consider migit be the following layout
Region) (SubjectMatters)\py mm) edith
Community vote dstrbuton
© & sagga TERTIETS) ye, months ago
tp docs microsoft conen-u/szure/storage/blob/ dat ake-strage-best-practicestbatchobs-structure
pte 51 tines
© & Deekshat234(Toxtoani5) 3 months, 2 weeks ego
correct,
pte Times
1B & examtopicseap 3 months, 3 weeks ago
Discorrect
‘pte T nas
2 & StudentFromAus § months 250
Comect
‘puted T nes
BM Doth 6 months, 3 weeks ago
Dis correct
pte Times
© Olukunmi 7 months, 1 week ag
Dis correct
‘pte Times
© & Egocentie 7 months, 2 weeks ago
Dis correc
pte T ines
© & SebK 8 months, 2 weeks 260
Dis cone
pote tines© & Ralphtiang 9 months ago
Discorrect
‘pte T nes
© NeeralKumar 9 months 290
Comect
‘pote T nes
© & PallaiPatel 10 montis, 1 week 290
Comect
‘puted Times
2 & Skyrocket 10 months, 2 weeks 290
‘pte T tines
2B Verodon 11 months, 1 week ago
Thats comet
‘puted 7 tnes
©. Mahesh.mm 1 months," week age
pte T nee
© & slexleonvaleneia 1
Respuesta correcta D.
pte Times
2M rashjan 11 months, 4 weoks 299
Dis corre
pte Times
© & ohana yer, 1 month sg0
Took the exam today, this question came out
Ans: D
‘pve neeHoTsPor
You need to output fles from Azure Data Factory.
Which file format should you use fo each type of output? To answer, select the appropiate options inthe answer area,
NOTE: Each correct selection I worth one pont
Hot ea:
Answer Area
Columnar format: v
Avro
Gzip
Parquet
TT
JSON with a timestamp: v
Avro
Gzip
Parquet
TXxT
Answer Area
Columnar format:
Avro
GZip
Parquet
Correct Answer: TT
JSON with a timestamp:
Avro
Gzip
Parquet
TxT
Box: Parquet
Parquet stores datain columns, while Avro stores data in a row-based format. By their very nature, colummorented date stores are optimized
for read-heavy analytical workloads, while row-based databases are best for write heavy transactional workloads
Box2: Avro
‘an Avo schema is crested using JSON format.
AVRO supports timestamps.
Note: Azure Data Factory supports te following fle formats (not GZip or TXT).
Avro format
~ Binary format
= Delimited text format
> Excel format
JSON format
=> ORC format
“> Parque format> XML format
Reference:
btps:rwww.éatanami.com/2018/05/6/bigdatefile-formats-demystiied
© & Matesh, mm ([GETEEE) 1 mons, 1 week a90
Parquet and AVRO i corect option
peed 26 tines
© & sppd 2 months, 3 weeks 290
Parquet and AVRO is correct
pte T ives
© A Deckshat234 3 months, weeks 259
‘pte Times
2M Razat2 4 months, 2 weeks 290
Parquet and AVRO is correct option
upvted ines
BA Doth 6 months, 3 weeks 90
agree withthe answer
‘puted 3 mas
© & Ratphtiang 8 months ago
Parquet and AVRO is correct option
pvted Zines
© a Pala
conrect,
pte Times
tel 10 months, 1 eek 290
1B & Shyrocket 10 months, 2 weeks ago
Parquet and AVRO is right.
puted? ines
2 & dba 11 months, 1 week ago
CGP fle format is one of supported Binary format by ADF
https/docs microsoft com/en-us/azure/data-factory/cannector-fle-systemttabs = data-factorytfle-system-35-sink
puted Tinas
2d bad atitude 21 months, 3 weeks ago
agree withthe answer
upted tines
© & alextoonvalencia 11 manhs, 3 weoks 290
Respuesta correcta PARQUET & AVRO.
pte TimesHoTsPor
You use Azure Data Factory to prepare data to be queried by Azure Synapse Analytics serverless SQL pools.
Files are iniallyngeste into an Azure Data Lake Storage Gen? account as 10 small SON files. Each fle contains the same data attributes and
data from a subsidiary of your company.
You need to move the files toa diferent folder and transform the data to meet the following requirements
* Provide te fastest possible query times.
"Automatically infer the schema from the undetying les.
How should you configure the Data Factory copy activity? To answer, select the apotootate options inthe answer area
NOTE: Each correct selections worth one pont,
Hot Aven
Answer Area
Copy behavior: wi
Flatten hierarchy
Merge files
Preserve hierarchy
Sinkfiletype: | SSsi
csv
JSON
Parquet
TXT.
Answer Area
Copy behavior: wi
Flatten hierarchy
Merge files
Sink file type S
csv
JSON
Parquet
TXT.
Box 1: Preserver hierarchy
‘Compared to the at namespace on Blab storage, the hierarchical namespace greatly improves the performance af directory management
operations, which improves overall job performance
Box2: Parquet
‘Azute Data Factory parquet format is supported for Azure Data Lake Storage Ger2
Parquet supports the schema propery
Reference:
hitpsffdocs microsoft. com/en-us/azure/storage/blobs/ data lake storage introduction htps://docs micrasoft.com/en-us/szure/data
factory/format parquet
© & alana GTTTTS1 yeay, 6 months ago
{Merge Files
2. Parquet
Iitps//docs microsoft con/en-us/azure/storage/blobs/data-ake-storage-performance-tuning-guidance
‘poted 17 tes© & Ameenymous 1 yes: 6 months age
‘The smaller the files, the negative the performance so Merge and Parquet seems tobe the right answer.
upoted 16 times
2 A edba 11 months, 1 wesk ago
just want to add a bit more reference regarding copyBehavior in ADF pls info mentioned in Best Practice doc. soit shall be Mergfil first
Ittps/docs mieosot com/en-us/azure/data-factory/cannector-fle-systemttabs= data-factoryefle-system-35-sink
ted mee
1B & lowd 1 year, 5 month ago
Larger fies lead to better performance and recuced costs
Typically, analytics engines such as HDInsight have a per-file overhead tht involves tasks such 3 litng, checking access, and performing
various metadata operations. you store your data 3s many small les, this can negatively affect performance. In general, organize your data
into larger sized files for beter performance (256 MB to 100 GB in size). S
poted times
© & captanbee (GETS)! yew, Smorts ago
Its frustrating just how many questions Examples get wrong Cant be helpful
pveed tines
© & gssdascoder* year, 1 month ago
Trying ta understand ifn answer iscorect will hep learn more
puted 7 ines
© & RywHayabusa 1 yea 4 mors ago
‘At east helps in learning, as you have te research and think for yourself Another big topics having tis questions inthe fist place is
immensely helpful
voted «2ines
© & Maviodiagps S months, 3 weeks 200
itis missleading, they should not give any answer at al
\pvted2imet
© & Sebk 8 months, 1 week ago
Agree,
‘ted 2 ines
© & Selmas7 TE) weeks, 1 ay ogo
| think "Atomatcallynfrthe schema from the underlying fle” means we should keep the same hierarchy and not merge all the data into 2
single fe. So I would say thatthe frst one is Preserve Hearchy.
‘pte T nes
PA Oldschool 2 weeks eso
‘Ql are intially ingested into an Azure Data Lake Storage Genz account as 10 small JSON files. Each file contains the same data attributes
and data from a subsidiary of your compary.
Since all have same data and attnbutes we ean Merge them in one file and automaticaly infer the schema from the underbing 10 smal fies.
Merge
Parquet
ited nee
© & Azure Kraken 3 months, 1 wee 390
‘So many answer are misleading, totally fed-up by checking the comment section far each and every question
‘puted 7 tines
2 W tomas 1 month, 2 weeks ago
This sit isn't a brain dump, is 9 study tool and it very often has some insights that are actually useful in my job beyond the tests I all you
want is answers im sure that you can find something else tat fits the bil
‘voted T nes
© A tomras 1 month, 2 weeks ago
It's more along the lines of being 2 Stackoverlow for the Microsoft exams.
‘ted net
© & Deckshat234 3 months, 2 weeks ago
1. Merge Files
2 Parquet
‘puted T nas
© id months, 1 wack ago
| would say Merge might not be possible, ecause we do nat know i the Json contains enough information about the subsidiary, which we would
need ifwe merge
pte Times
BB Atha months 992
The frst answer should be Merge Files‘pte inas
2 & azures0otest 5 mons 999
1. Merge Files
2. JSON
Decompression requires compute and thus costs time, Since the ls are small eavng them uncompressed is ok since it will ll faster query.
‘Also they are JSON ies so compressing them in colurmar format does not make much sens,
puted ines
2 & Dathy 6 months, 3 weeks 290
1. Merge Files
2 Parquet
‘puted 2 mes
© & KashRaynardMorse ? mots, 3 weeks 990
‘Atequirement was "Automatically infer the schema from the underlying fles", meaning Preserve hiratchy is needed
‘pvted 3 nas
oa
Tonite 290
is possible that all or some schemas may be ifferent in that case we cannot merge
voted I nee
1B & Aditya049 5 mans, 2 weeks 290
gabdu why are you tying to create a diferent use cas? it's clearly mentioned thatthe data attributes are same in all the files. Answers
merge, don't canfuse others looking for answers nere
‘pote tines
© & imomins& months ago
another hat key is: You need to move the files toa diferent folder
0 answer should be preserve hierarchy.
‘puted? tnes
BM adizra54 2 mons, 1 week 250
But the data attributes ave same inal the ies
pote Tes
2 & Eyepateh99a e mont, 1 week ago
1. Preserve elrarchy - ADF is used only for processing and Synapse isthe sink. Since synapse has parallel processing power, it can process thefls
in diferent flder and thus improve performance.
2 Parquet
‘puted Times
© & uzairah 5 montis, week ago
‘The question says "Provide the fastest posible query times." not the wite times an there is na in rectory there are just 10 json files so
combining them into one file would be most approprite a large fl size increase performance rea iy link
https//docs microsoft com/en-us/azure/storage/olobs/data lae-storage-best-praclcestfile-size
voted T nes
2 & kami 2 months, 2 weeks 290
‘Ae these answers the actual correct answers or guesses? Who highlights the correct answers?
pte nes
BM sraken 10 months, 1 week go
“In general, we recommend that your system have some sort of process to aggregate small les into larger ones for use by downstream
applications.
https//docs microsoft com/en-us/azurefstorage/olabs/datalake-storage-best-practices
‘Thetefore I believe the answer should be "Merge fles'and Parquet’
‘pte tines
1B & Sandipau 10 months, 3 weeks 290
Merge ané parquet willbe the right option, also taken reference from Udemy
pted2simes
© Mahesh sm 11 months,“ week 390
“LAs hierarchical namespace greatly improves the performance of directory management operations, which improves overall job performance,
Preserverherarchy looks correc. Also there is overhead for merging les,
2 Parquet
‘puted 3imas
© & Boompiee s months, 3 weeks apo
‘The overhead for merging happens once, after that it’ faster every time to query the files if they are merges.
pote times© & urairahm S montis, week 250
‘The question says "Prove the fastest posible query times." not the ite times and there is no in dlvectory there are just 10js0n files so
combining them int one file would be most appropriate as large fl size increase performance read in link
https//decs microsoft com/en-us/azure/storage/olobs/data lae-storage-best-practcestfle-size
ted nee
© & mashines 11 months, 3 weeks ago
Merge Files and Parquet
‘puted tinesHoTsPor
You have a data model that you plan to implement in data warehouse in Azure Synapse Analytics as shown inthe following exhibit.
Fact_DailyBookings Dim Customer
PERE iDailyBookingsID iCustomerlD
iEmployeelD. iCustomerlD >———#_ vcCustomerName
vcEmployeeLastName iTimelD vcCustomerAddress1
vcEmployeeMName iEmployeetD veCustomerCity
veEmployeeFirstName fp ——oF ttemID
atEmployeetireDate iQuantityOrdered =
dtémployeetevel dxchangeRate Dim_Time
dtEmployeeLastPromotion iCountryofOrigin iTimelD
- mUnitPrice + icalendarday
iCalendarWeek
iCalendarMonth
veDayofWeek
veDayofMonth
veDayofYear
Azure Synapse Analytics iHolidayindicator
‘Allthe dimension tales willbe less than 2 GB after compression and the fact table willbe approximately 6 TB. The dimension tables willbe
relatively static with very few data inserts and updates.
Which typeof table should you use fr each table? To answer, select the appropriate options inthe answer area
NOTE: Each correct selection I worth one pont.
Hot Area:
Answer Area
Dim_Customer: v.
Hash distributed
Round-robin
Replicated
Dim_Employee: Vv
Hash distributed
Round-robin
Repiicated
Dim_Time: v
Hash distributed
Round-robin
Replicated
Fact_DailyBookings v
Hash distributed
Round-robin
ReplicatedAnswer Area
Dim_Customer: v
Hash distributed
Round-robin
Replicated
Dim_Employee: v
Hash distributed
Round-robin:
Correct Aner: Epica!
Dim_Time vy
Hash distributed
Round-robin
Replicated
Fact_DailyBookings: v
Hash distributed
Round-robin
Replicated
Box 1: Replicated
Replicated tables are ideal for small starschema dimension tables, because the fact table is often distributed ona column that's not
compatible with the connected dimension tables. this case applies to your schema, consider changing small dimension tables currently
Implemented as round:obin to replicated
Box 2: Replicated
Box 3: Replicated
Box 4: Hash-distbuted
For Fact tables use hash-distribution wit clustered columnstoe index. Performance improves when two hash tables are joined onthe same
Aistibution column
Reference:
hitpsfazure microsoft. conv/en-us/updates/reduce datz movement-and-make your queries more-fclent withthe general-avallablty of
replicated tabes/htips:azure microsoft com/er-us/blogireplicatedtables-now-generall-avalablein-azure-sqkdta.warehouse/
© Bin viana TET) yea, 2 monte ogo
‘The answer's correc.
‘The Dime are under 2ob so n0 point in use hah,
‘Common distribution methods for tables:
‘The table category often determines which option to choose for distributing the table
Table category Recommended distribution option
Fact -Use hash-distibution with clustered columnstore index, Performance improves when two hash tables are joined on the same distribution
colar
Dimension - Use replicate for smaller tables. fables are too large to store on each Compute node, use hash-dstrbuted
Staging - Use round-robin forthe staging table. The load with CIAS i fast. Once the data i inthe staging table, use INSERT. SELECT to move the
data to production tables.
htps//docs microsoft com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-data-warehouse-tables-overviewtcommon-distribution-
methods-for-ables
puted 11 tines
© virendrapsingh § montis, 3 weeks 290
This is a wonderful explanation, Worth giving ike
‘voted 3 ines
© Gametie yea, 1 month ago
Thanks, but where in the question does it indicate about Fact table has chisteredcolurmnstore index?‘puted 3 ines
8
rserksap 1 yea, month 299
Normally for big tables we use casted colurmnstore index for optimal performance and compression, Since the table mentioned heres in
TBs we can safely assume using this index is the best choice
https//docs microsoft comy/en-us/azure/synapse-analytics/sol-dat-warehouse/sel-data-warehouse-tables-index
pated 2imet
© & berserksap year, 1 month ago
https//docs microsoft com/en-us/azure/synapse-analytics/sq/develop-tables-overview
‘ted nee
© & ohana (GETS) 1 ye, 1 mam 290
Took the exam today, this question came out
‘Ans: All the Dim tables > Replicated
Fact Tables -> Hash Distributed
pred 3 tines
© & Deekshat234 (Terme) 3 moths, 2 weeks 290
contect
poted 1 tines
B & bjecto 5 months ago
Just a better link that explains the decisions. Also watch te video, its cool
https/docs mirosote com/en-us/azure/synapse-analytics/sql-data-warehouse/massivly-parallel-processing-mpp-architecure
puted 2times
© & Dothy 6 months, 3 wesks 290
‘The answers correct,
puted Times
© & Pallavipatl 10 montis, 1 week 290
upted times
© & Pritamts 10 months, 1 week aso
Got this question on 2312/2021. answer is core
pte ines
© A Mahesh sm 11 months," week 390
Ansis correct
upted ines
© alfonsedisatve 1 yea, 1 month apo
Dimension are Replicated
“Since the table nas multiple copies, replicated tables work best when the table sze sles than 2 G8 compressed,
“Replicate tables may not yiela the best query performance when
‘The table has frequent insert update, and delete operations"
= We recommend using replicated tables instead of round-robin tables in most cases"
https//docs microsoft com/en-us/azure/symapse-analyis/sql-data-warehouse/design-guidance-for-replcated-tables
posted Times
BA gssdéscoder * year 1 morth ego
Comect:httos//decs microsoft com/en-us/azure/synapse-analytis/sq/develop-tables-overview#common-distrbution-methodsfor-tables
pute timesHoTsPor
You have an Azure Data Lake Storage Gen? container.
Data ingested into the container, and then transformed by a data integration aplication. The data is NOT modified after that. Users can read
Sle nthe container bu cannot modify the files,
You need to design a data archiving solution that meets the following requirements:
= New datas accessed frequently and must be available as quickly as possible,
~» Data that solder than ve years Is accessed infrequently but must be avallable within one second when requested.
® Data tats older than seven yeas is NOT accessed. After seven years, the data mustbe persisted at the lowest cost possible.
"> Costs must be minimized while maintalning the required availabilty
How should you manage the data? To answer, select the appropriate options inthe answer ara,
NOTE: Each correct selections worth one point
Hot Area
Answer Area
Five-year-old data:
Delete the blob,
Move to archive storage.
Move to cool storage.
Move to hot storage.
Seven-year-old data:
Delete the blob,
Move to archive storage
Move to coal storage.
Move to hot storage,
Answer Area
Five-year-old data:
Delete the blob,
Move to archive storage.
Move to cool storage.
Correct Answer: | Move to hot storage.
Seven-year-old data:
Delete the blob,
Move to archive storage.
Move to coal storage.
Move to hot storage,
Box 1: Move to cool storage
Box 2: Move to archive storage
Archive - Optimized for storing data that i rarely accessed and stored for at least 180 days with Mlexible latency cequirements, onthe order of
hours.
Te following table shows a compatison of premium performance block blob storage, andthe hot, cool, and archive acess ters,Availability
Availablity
(RA-GRS
reads)
Usage
charges
Minimum
storage
duration
Latency
Time to
first byte)
Reference:
hitpsifdocs microsoft. com/enus/azure/storage/blobsstorage-blobstoragetiers
Premium
performance
oe9%
NA
Higher storage
costs, lower
access, and
transaction
cost
NA
Single-ciait
riliseconds
Hot ter
pag
92998
Higher storage
costs lower
access, and
transaction
costs
NA
rmiliseconds
© & yobhip Gaara) year, 6 months ago
Answer should be
17 Coal
2 Archive
oo! tier
208%
29.0%
Tower storage
costs, higher
access, and
transaction
costs
30days?
‘Comparison table shown access time for coal tert is milseconds
Archive tier
oftine
oftine
Lowest storage
costs highest
access, and
transaction
costs
190 days
hutpsi/docs mirosof com/en-us/azure/storage/blabs/storage-blob-storage-tersttcomparing-black-blob-storage-options
puted 68 tines
2 & r00s 6 months, 1 week 290
Right #1 is Cool because its clearly mentioned in the dacumentation that “Older datasets th
available for immediate access
https/does mieroso% com/en-us/azure/storage/blobs/access-ters-overiew#comparing-blockcblob-storage-options
ted nee
© & sustbu (GETTEETS)1 yen, 2 months 290
“Tricky question, t says data that is OLDER THAN (> 5 years), must be available within one second wen requested
not used frequently, but are expected to be
But the frst question ass for Five-year-old data, which s=5, 50 It can also be hot storage
Similarly forthe seven-year-old
Not sure, please
puted 1 times
sonfiem?
BB joshtey 4 months ago
“Costs must be minimized while maintaining the required availabilty,
pote times
© & PugazhManohar[Sa=a==1) 3 months, 3 weeks ag0
cod, 2.Aschive
puted Times
© examtopicseap 3 months, 3 weeks ago
‘Theres @ question with ‘data over 5/7 yeas old so it's very tricky
1-Hot
2-cold
pte times
2 & Dothy 6 months,
3 weeks 390
So cold and archive ate the correct answers,pte T ines
2 & Pallavipatel 10 montis, 1 week e90
pte Ties
BB ANath 11 ments ago
1. Cool Storage
2. Archive Storage
pte Times
© & Mahesh.ymm 1 months“ week 290
Answer is corect
puted Times
2M sith year montis ago
[Answer should be
thot
2srehive
https//mmwbme com/blogs/cold-vschot-datastorage/
Cold storage data retrieval can take much longer than hot storage. can take minutes to hours to access cold storage dats
puted times
© & syamumar 1 year S months 350
[also doubt fits hot storage and archive. because its mentioned S-year-old has to be retieved within seconds which snot possible via cold
storageli
ited Tee
© & savin yar $ months ago
but the cost factar i also there. keeping the data in hotter for S years vs cod tie for 5 years would add signfiant amount
‘pote T nes
© & captainbee + yea, § months ago
old storage takes aillseconds to retrieve
pte tines
© & mareint212 1 monthe,2 weeks ag0
httpsi/wnwbmecom/blags/eold-vs-hot-data-storage/
Riso about Azure!
ted 2 inet
2 & Dre 1 year 6 months 90
‘Answer is corect
pted inesQuestion #11 opie 1
RAG DROP.
You need to create a partioned table in an Azure Synapse Analytics dedicated SQL pool
How should you complete the TrensactSQL statement? To answer, crag the appropriate values tothe corect targets, Each value may be used
‘once, more than once, o nat at all You may need to drag the spit bar between panes ot sral to view content,
NOTE: Each correct selections worth one pont
Select and Place:
Values ‘Answer Area
CLUSTERED INDEX CREATE TABLE tablet
= ID INTEGER,
DISTRIBUTION colt VARCHAR (10),
PARTITION col? VARCHAR (10)
PARTITION FUNCTION ; ae
Pesci ees =f ") = HASH(ID),
(a sacs uber FoR VALNES (1, 1900000, 2000000)
¥
Corest Answer:
Values ‘Answer Area
CREATE TABLE tablet
“Gg seca,
[Dremerporron | col VARCHAR(20),
[parrerron—] __coll2. varcxar(io)
EARTITION FONCTION)» ®T7H
DISTRIBUTION HASH (TD)
(20 RANGE LEET FoR VALUES (2, 1000000, 2000000))
b
Bo 1: osTaBUTION
‘Table distribution options include DISTRIBUTION = HASH { distribution column_name ), assigns each row to one distribution by hashing the
vali stored in dition cou nae.
Box 2; PARTITION
Table partition options. Syntax:
PARTITION (patton column same RANGE [LEFT | RIGHT] FOR VALUES (| boundary.value [1] })
Reference:
htpsildocs.microsoftcom/enus/sqlt-sq/statements/createtable-azure-sqhdata-warehouse
© & Sunny TE
poise
© & Sasha in San Francisco (STS) 1 yea ago
Correct answer by haw to remember? Distribution option before the Partition option because.."D" comes before” or because the syste needs
‘to know the algorithm (hash, round-rbin, replicate) before it can stat to Partition o segment the data, (seem reasonable?)
puted $0 times
© & Deekshat 224 EATS) 3 months, 2 weeks ago
correct
preted nes
x 6 months ago
© & gursimran_s § months, 4 weeks ago
6o with a logical explanation guys. what i this D before P.if take i ike that then C comes before Das well. Try to grasp the logics. answers
conrect.
‘pted times
BA Dotty & months, 3 weeks ago
Answer is corect
puted times© A Egocentic 7 months, 2 weeks ago
provided answer is corect
‘pte Times
© A vineeti234 8 montns ago
D comes before Pas in DP-203
‘puted 7 mes
© & PallaviPatl 10 montis, 1 week 290
puted times
2 & Jaws1990 11 montis 290
‘Wouldt VALUES(, 1000000, 200000) create a partition for records with 1D
pte Times
BM ploer 9 months, 4 weeks ago
Having three boundaries wil lead to four pattions
1. Parttion for values <1
2. Partition for values from 7 to 999999
3. Partition for values from 1000000 to 199999
4 Parton for values »= 2000000
puted 7 tines
© nastyana 9 montis, 2 weeks 290
but only <= and >. tis range let for values right
pote tines
© & Mahesh.mm 11 months," week 290
Answer is corect
pte T ines
© & hugoborda 1 yer 2 months ag0
Answer is corect
pte Times
© & hsetn 1 yes, 3 months 90
Indeed! Answaris correct
pte Times
which would mean 1 row?You need to design an Azure Synapse Analytics dedicated SQL poo that meets the following requirements:
‘> Can return an employee record from a given point in time.
-* Maintains the latest employee information,
Minimizes query complexity.
ow should you model the employee data?
‘Aas atemporal table
B.as SQL graph table
(as a degenerate dimension table
1. as a Type? slowly changing dimension (SCO) table
Correct Answer: D
‘AType2 SCD supports versioning of dimension members. Often the source system doesnt store versions, so the data warehouse load process
detects and manages changes in a dimension table. In this case, the dimension table must use surogate key to provide a unique reference to
aversion ofthe dimension member, als includes columns that define the date range validity ofthe version (fr example, StartDate and
EndDate) and possibly a fag column (for example,
IsCutent) to easly filter by current dimension members.
Reference:
tpsifdocs. microsoft. com/en-usylearn/modules/populate-slowly-changingdimensions-azure-synapse-analytis pipelines/3-choose-between-
dimension-ypes.
Community vote dstrbuton
© & pesaceses [Ham Ia=za) yeu, mons ago
‘Answer D: Temporal tbl is better than SCD2, but ts no supported in Sympase yet
preted tines
© & Preben + yess, Smonths ago
Hore’s the documentation for how to implement temporal tables in Synapse from 2018,
https//docs microsoft com/en-us/azure/synapse-analytis/sq-data-warehouse/sql-data-warehouse-tables-temporary
ted nee
© & mbravo 1 year,S montis 290
‘Temporal tables and Temporary tables are two very distinct concepts, Your link has absolutely nothing to co with thi question,
voted Tatines
© Vaishnav 1 year 4 montns ago
\psdocsmicrosof.comven-us/azure/azure-sq/temporal-tables
Answer :A Temporal Tables
‘voted ines
© berserksap 1 year, 1 month ago
| think synapse doesnt sup
‘voted nee
temporal tables, Pease check the below comment by hsetin
© & sparkohus mons, 3 week ago
though this not something relative to this question. temproal tables lock ale to deta table
coveted tines
© & rashan(HERTTERTS) 1 months, 4 weks ago
Dis correct (voting comment that people dont have to open discussion always, please upvote to help others)
puted 67 times
© & oldschool (ATED)? woos, 5 days 99
scb2-D
pote Times
© A Deekshat234 3 months, 2 weeks apoconreet
‘pted 2 nas
© & PugazhManohar 3 months, 3 weeks 200,
Maintains the latest employee informa
pte Times
= SCO-Type (Ans-D)
1B & Dathy 6 months, 3 weeks 290
Answer is corect
posted 2simes
© & Martn_Nbg 7 mons, 7 week 290
“Temporal tables ae not supported in Synapse so is correct
puted 7 ines
2 & sparkchu 8 months, 1 week ago
overall you should use delta table
‘ted T nes
‘puted T ines
© & Adelina 10 months, 3 weeks ago
Discarract
‘puted 2 nes
© & devadey 10 mors, 4 weeks 290
Confusing high voted comment. Ds SCD? but comments talking about temporal table Fier way SCD? right answer which is Choice D
‘puted Times
2B Veropon 11 months, 1 week ago
Dedicated SOL Pools is the key
‘puted 3 ines
© Maesh.mm 17 months," week age
Answer is D
pte Ties
© M hesetin 1 yea, 3 months age
‘Answer i D, Microsoft seems to have confirmed ths.
https//docs microsoft. com/en-us/answers/questions/13056 /temporal-table-n-azur
synapse htmit~text= Unfortunately22C200we%20do%20nctX20supportsubmitted’k20by'20another'é20Arure%20customer.
‘pte nee
© & dav722 5 year 3 months ago
‘Answer D is correct Temporal tables mentioned inthe lnk below are supported in Azure SQL Database(Paas) and Azure Managed Instance, where
2 in this question Dedicated SQ1 Pools are mentioned so no temparal tables can be used, SCD Type 2 the answer
hpsi/docs mirosof comven-us/azure/azure-sql/temporal-tables
puted 5 tines
(© @ exeoins 1 year, Smonths ago
Definitely answer D
pted 3 times
© & [Removed] ea, s montis a2
‘The answer is A = Temporal tables
"Temporal ‘ables enable you to restore row versions from any point in time
https//decs microsoft com/en-us/azure/azure-sql/database/business-contiuity-high-aallabilty-isaster-recover-hadr-overview
puted Timesauestion #13 opie 1
You have an enterprise wide Azure Data Lake Storage Gen2 account. The data lake is accessible ony through an Azure virtual network named
wert
You are building @ SQL poo in Azure Synapse that will se data from the data lake,
Your company ha a sales team, Allthe members of the sales team are in an Azure Active Directry group named Sales, POSIX controls are used
to assign the
Sales group access tothe files inthe data lake,
You plan to lad data to the SQL pool every hour
You need to ensure that the SQL pool can load the sales data from the data lake.
Which three actions should you perform? Each correct answer presents part ofthe solution
NOTE: Each area selection is worth one point.
‘A. Add the managed det tothe Sale group.
8. Use the manage dently a5 the credential forthe data load process
C. Create a shared access sgnatute (SAS).
©. Ad your Azure Active Directory (Azure AD) account tothe Sales group.
Use the shared acces signature (SAS) as the credemals forthe data load process.
F Create a managed identity
Correct Answer: BF
‘The managed identity grants permissions tothe dedicated SQL pools inthe workspace
Note: Managed identity for Azure resources isa feature of Azure Active Directory. The feature provides Azute services with an automatically
managed identity in
‘Azure AD-
Reference:
hitpsffdocs microsoft com/en.us/azure/synapse-analytcs/securty/synapse-workspace-managed-dentity
Community vote dsmbuton
© & diane (GETEI)1 ye 6 mons 290
correct answer is ABFhttps/www examtopiescom/dscussions/microsof/vew/1207-exam-dp-200-ropic-1-question-S6-discussion/
upoted 721s
© vith year smonths 290
Yes but the order is clffeent itis FAB
ted 35times
© A Kinglo 1 yea:, 3 months 290
‘The question did't speciy erder or sequence
ted 10ines
A gesdscoder * yoo,
Bracth, agree with you
pated nee
© A ikl (aT) 1 you 4 months ago
Comect Answer should be
F Create a managed identity.
{A dd the managed identity to the Sales group,
B Use the manages identity a5 the credentials forthe data load process
pvted 30 tines
© & ituox (ERT) week, 3 doy ago
month 390,
Fist create a managed ID, Add the managed ID, use the managed ID
pte Times
© & mahantal | month weske ag0
Can anyone explain why we cannot useshared access signature inthis case?‘pte inas
© & dmitrypo 1 ont, 1 week ago
“Managed Identity authentication is required when your storage account i attached to 8 VNet.
https: leotn microsoftcom/en-us/saute/synapse-analics/sq)-data-warehouse/quickstart-bulk-load-copy-tsal-examples
‘voted T nas
BB eval 3 months, 2 weeks ago
AGF isthe corract ane
pote Times
© A Deekshat234 3 months, weeks ago
ABE is conrect
pted times
© & Dethy 6 months, 2 weeks 290
correct answer is ABP
puted times
2 & Egocentie 7 months, 2 weeks ago
ABFis correct
‘puted T ines
© @ praticewizards 8 months," ek ago
FAB - create, add to group, use to load data
‘puted T tines
© & wacky 11 montis a90
Is answer A properly worded?
“Aad the managed tent to the Sales group” should be “Ad the Sales group to managed identity"
preted ines
© & lukeoatine 11 montis 90
FAB should be correct
pte ines
BM Verodon 11 months, 1 week ago
FAB ip correct sequence
upvted ines
© Sabadamal2010atGmall 11 months, 1 wesk ago
1. Create a managed identity
2, Add the managed identity tothe Sales group.
3. Use the managed identity as the credentials forthe data load process
pted times
1B & Maheshymm 1 months,“ week 290
FAB ie correct sequence
puted Times
© & Lewistrck 11 mons, 2 weeks ag0
‘Would it even be a good ides to have the data load process be part ofthe Sales team? They have separate responsibilites so should be part of
snather group. | know that’s not possible in the answer st, ut" tying to think best practices here
puted tines
2 & Astam208 1 yeo:392
Comect answer's FA, 8
puted tines
2 & FredNo 1 year a90
use managed identty
puted SinesQuestion #14
HoTsPoT
‘You have an Azure Sypapse Anabytics dedicated SOL poo! that contains the users shown inthe folowing table
Name Role
Usert___| Server admin
User2 | db_datereader
seri executes a query on the database, andthe query retuns the results showa inthe following exhibit.
ebl.name a2 table_nam
typ-nane as datatype,
4 cris masked,
FROM sys-maskod_columns AS ¢
INNER JOIN ays.tables AS tbl O¥ ¢. (ob:
INNER JOIN sys.types typ ON ¢.user_typ
WHERE ie masked = 07
Ad = typ.user_type_ia
Results Messages
name table_name datatype —_is_masked —_ masking_function
1 | BirthDate | DimCustomer date 1 default ()
2 | Gender DimCustomer | nvarchar | 1 default ()
3 | EmailAddress |DimCustomer nvarchar | 1 email ()
4 | YearlyIncome | DimCustomer | money 1 default ()
Usert isthe only user who has access to the unmasked data,
Use the drop-down menus to select the answer choice that completes each statement based on the information presented inthe graphic.
NOTE: Each correct selections worth one point
Hot ares:
Answer Area
When User2 queries the Yearlyincome column,
the values returned will be [answer choicel. v
a random number
the values stored in the database
XXXX
0
When Userl queries the BirthDate column, the
values returned will be [answer choice]
a random date
the values stored in the database
XXX
1900-01-01Correct Answer:
Answer Area
‘When User? queries the Yearlyincome column,
the values returned will be [answer choice] v
a random number
the values stored in the database
XXX
0
When Userl queries the BirthDate column, the
values returned will be [answer choice]. v
a random date
‘the values stored in the database
00K
1900-01-01
Box 1:0
The Yearlyincame column sof the money datatype
‘The Default masking function: Ful masking according to the datatypes ofthe designated fields
» Use a zero valve fr numeric datatypes (isin, bit, decimal int, money, numeric, small, smallmoney,tnyn, loa, eal),
Box 2: the values stored nthe database
Users wth administrator privileges ae always excluded from masking, and see the orginal data without any mask.
Reference:
btpsf/docs.mierosoft.com/en-us/azure/azure'sal/database/dynamic-date-masking overview
© & hseta (ERT TOE)1 yea, 3morths ago
User Tis admin, so he wl see the value stored in dbms
10
2. Value in database
puted 65 times
© & anurearmy year, 1 month 290
Bis wrong
voted? nes
2
‘Aaltya0891 Sons, 2weoks ago
zureammy read the question properly anl then answer, nd is queried by user, masking doesnt apply to user?
‘pote tines
© & ile Gama year, 3 months ago
+ Use a zero value fr numeric dats types (bignt. bit, decimal, nt, money, numeric small. smallmaney,tinyint, fot rea
+ Use 01-01-1900 for datetime data types (date, datetime2, datetime, datetimeofiset,smalldatetime, time).
prted T7imes
© & berserkeap 1 yea: 1 month ago
‘The second question is queried by User 1 who is the acmin
ted 7 times
© & ovésehool (Sn) wees, days ag0
User is reader 0 he wl 00 querying Yearhincame with defauk mask;
User is admin and ony he wil seal stored values
preted nes
© A Deckshat234:3 months, 2 weeks 250
‘pvted 3 nas
© & azures0otest 5 months 299
User 1 The value
User 2.100%
see https//docs microsoft com/en-us/szure/szure-sal/database/dynamic-data-masking-overview?view=32uresqlAND
https/wwselshack com/ynamie-data-masking-n-sol-server
pte 2 ines
BW objecto 5 months, 3 weeks ago
According to https//docs microsoft comven-us/sl/elational-databases/security/dynamic-data-maskngview=sol-server-ver16
“For date and time datatypes use 01.01.1900 0:00:00.0000000 (dat, datetime2, datetime, datetimeottset,smalldatetime, time),
Data masking as defau
puted Times
1a date dateime smalldate should be 1900-01-01. Strangely here is no such option Any ideas anyone?
© & objecto 5 months, 3 weeks 290
Damn, User 1 only reads Yearlylncome (not date, 50 yes Os the corect answer
poted 7 ines
2 & Doth & months, 3 weeks ago
10
2, Value in database
‘pote Times
BB Egocentie 7 months, 2 weeks ago
con this question its just about paying attention to detail
‘puted Sines
© & manant6 7 montis, 4 weeks ago
How user? can access data as itis masked?
puted times
© & manant6 7 months, « weaks ago
Can Someone explain frst option asin doc it says 0
‘puted T nes
© Mahesh.mm 1 months," week age
1.0 (Default values for money datatype for masked function will writen when queried by user2)
2. Value in database (As it is queried by user! whois admin)
‘pted 8 times
2 Milan
correct
‘poted times
1 yeat,1 month ago
BB gssdéscoder year 1 month apo
‘Agree with answer, but | 38 a typo in the question db_datereader MUST be db_ datareader,
puted 3 mes
2 & du 1 yea, 1 month ago
0 for money and 1/1/1900 for dates
https//docs microsoft com/en-us/sql/elationa-databases/security/dynamic-data-masking?view=sel-server-ver1S
pote 5 tines
© & ditya049r 5 mons, 4 wesks ago
attention to detail, What you mentioned is the default used for date incase of masking but here the query is executed by user 1 who has
‘access othe table and no masking rues will be applied for user 1 so that answer is value present in the database
ted 3 nes
© & GervasioMontaNelas 1 year, 2 montis 290
Its carect
puted? ines
Bolles year, 3 months ago
upvted tines
© Mazazing 1 yer, month 290
yes, its correct
ted 2imetYou have an enterprise data warehouse in Azure Synapse Analytics.
Using PolyBase, you create an external table named [Ext Items] to query Paquet files stored in Azure Data Lake Storage Gen? without importing
the data to the data warehouse.
‘The extemal table has three columns,
You discover thatthe Parquet files have a fourth column named ItemD,
Which command should you run to ad the ItemD column to the extemal able?
A
ALTER ETERNAL TABLE [Fat] . (Ttems]
‘ADD [ItemzD] ints
3
DROP EXTESNAL FILE FORMAD parquevfilel;
CREATE EXTERNAL FILE FORMAT parquetfilel
wrTE (
FORMAT_2YPE = PARQUET,
DATA_COMPRESSION = ‘org-apachs.hadcop.io.ccmpress.snappycodac’
io
«
DROP EXTERNAL TABLE [Ext]. [Ttems]
CREATE EXTERNAL TABLE [Ext]. [ztems]
({rtemrD] [int] woct,
[rtemvane} nvarchar (50) uz,
[itemtype]_nvarchar (20) NULL,
[Trempeseription] nvarchar(250))
wim
ci
LocaPION= ‘/rtems/",
DATA SOURCE = AzureDataLakestore,
FILE_FORMAT = PARQUER,
REJECT_TYPE = VALUE,
REGECT_VALUE = 0
”
0.
ALTER TABLE [Ext]. [Teens]
ADD [ItemzD] ints
Correct Answer: C
Incorrect Answers
‘A.D: Only these Data Definition Language (ODL) statements are allowed on exteral tables:
“> OREATE TABLE and DROP TABLE
> CREATE STATISTICS and DROP STATISTICS
“> OREATE VIEW and DROP VIEW
Reference:
bupsfdocs.mlerosoft.com/en-us/sql-sl/statements/createextemaltabletransact-sl
© & chien_Nguyen.Van TITIES) yor 3 months ago
Ciscorect
tps fmatexamtopice com dlscussonsmicrosofvew/19469-eram-dp-200-opic
vee tines,
© & oésehool (aaa) weeks, Says ag0
Cie Comect
pated nes
question 27-dscussion/
© & Selmas7 3 weoks, 1 ¢ay 290
[stil can understand why it's not
pte tines
© A Deckshat234 3 months, weeks ag0
correct,
pte T ines
2 & sp17 4 months, 3 weeks agoCis comrect. ven f you are confuse with other options. The cue here i Keyword Location while creating extemal able, LOCATION =
“older_or-flepath’ Species the folder or the ile path an file name forthe actual data
poted 2 tne
BW Oxren8 months, 2 weeks ago
Good thing the detail re shown here: “The external table has three columns." And the solution yet reveals the column details This doesnt make
‘any sense to me. If Cis the correct answer (only one that seems acceptable, then the question isl flawed
‘puted 2imes
BB dduquer 3 ys, 18 hours 90
‘The external table has 3 columins but the fie it references has 4 columns a the external table has to be altered
pte times
© & PallaniPatel 10 montis, 1 week a90
ciscomect
‘puted Tinas
© & hugoborda 1 yea, 2 mont ago
Answer is corect
pte tinesHoTsPor
You have two Azure Storage accounts named Storaget and Storage2. Each account holds one container and has the hierarchical namespace
‘enabled, The system has files that contain data stored inthe Apache Parquet format
You need to copy folders and files from Storage to Storage? by using a Data Factory copy activity The solution must meet the following
requirements:
“> No transformations must be performed.
.» The rignal folder structure must be retained
- Minimize time required to perform the copy activity
How should you configure the copy activity? To answer, select the appropiate options inthe answer are,
NOTE: Each correct selections worth one pont.
Hot Aes:
Answer Area
Source dataset type: lv
[Binary
Parquet
[Delimited text
Copy activity copy behavior: lv
[FlattenFierarchy
MergeFiles
PreserveHierarchy
Answer Area
Source dataset type: lv
Binary
Correct Answer:
[Delimited text
Copy activity copy behavior: lv
[FlattenFierarchy
MergeFiles
jerarchy
Box 1: Parquet
For Parquet datasets, the type property of the copy acivy source must beset to ParquetSource
Box2 Preservelirarchy
PreserveHierrchy (default: Preserves the file hierarchy in the target folder. The relative path ofthe source fete the source folder is identical
to the relative path of he target fle to the target folder
Incoreect Answers
-> FlattenHierarchy ll files from the source folder arin the fist level ofthe target folder, The target files have autogenerated names,
- Merges: Merges all les frm the source folder to one file Ith fle name is specie, the merged file name I the specified name,
Otherwise isan autogenerated fle name.
Reference:
hntpsffdoes.mierosoft.com/en-us/azute/dat-facory/format parquet htis:/éocs.mictosoft.com/envus/azure/dateTactory/eomector-azure-
datalake storage
© & eddyRoboto (Hanan) yor 9 monte ago
This could be binary a source and sink, since there are no transformations on files, tend to believe that would be binary the correct anwer.pvted 60 tines
BA 100) 9 months, 2 weeks ago
‘Agree, Ive checked it. With binay source and sink datasets it works
‘ted 2ines
© Gametite yer, 1 month ago
But the doc says “When using Binary dataset in copy activity, you can only copy rom Binary dataset to Binary dataset” So | guess it's parquet
then?
ted Stinet
© A captainpike 1 year, 1 month ago
‘This note is referting tothe fact that, inthe template, you have to specify "BnarySink” a the type forthe target Sink and that exactly what
‘the Copy data tool does. (you can check tis by eating the created copy pipeline and see the code) Choosing Slnary and PreserveHierarchy
copy al fil as they are perfectly.
poted 3 tines
2 Jed.ethak 1 year, 1 month ago
no it must be parquet because The type property of the dataset must beset to Binary and its parquet hear so answer are correct
‘voted T nas
© A michas 1 yet, 2 months ago
agree. If its just copying then binary is fine and would probably be faster
‘ated tines
© & AbhiGate Gara) ye, 2 mons ago
“Answer seems corect as data is store is parquet already and requirement isto do no transformation so answers right
protest tnes
© & NintyFour 6 months, 2 weeks 290
’As question hss mentioned, Minimize time required to perform the copy activi.
‘And binary faster than Parquet. Hence, inay is answer
pvted 3 mas
© & poko7 (TEER) days, hous ago
Answers comect,
Upveted tines
A temace 2 weeks, 2 daye 050
Binary - copy fles sis in fastest ay.
PreserveHicrarchy - for saving folder structure
pte Times
BA Oldschool 2 weeks, s days 290
“Answer i correct, No transformation and preserve hierarchy
puted times
© & RBKasemodel2 weeks, 5 days 399
[believe the answer shouldbe Binary, since its stated that no transformations must be done.
“You can use Binary dataset in Copy activity, GetMetadata activity, or Delete activity. When using Binary dataset, the service doesnot parse fle
content but eat asi
https//learn microsoft comyen-us/szure/data-factory/format-binary
| could’ found any information saying that parquet won't be parsed if the source and sink are parques fs. Sothink it will parse, and we can
Uungerstan that fis transformation
puted times
2 & allagowt 2 montis 290
Answer seems cortec,
advice don’ overthink the source is parquet and it's one ofthe options so iis parquet.
pote 3 tines
1D A Deekshat234 3 months, 2 weeks a50
siven ansis correct
‘pte Times
BA AKCI1 Amon, 2 weeks 250
Binary would work But since the question refer to the source filetype as Parquet and thats one ofthe option provided, | wauld go with Parquet
puted Times
© & RIKOT 6 months, 1 week 290
Agree with answer
puted times
© & NintyFour & months, 2 weeks 250{As question has mentioned, Minimize time required to perform the copy activi.
‘And binary i faster than Parquet. Hence, inary is answer
pte 2 tines
© & AnureRan 6 months, 2 weeks 90
Isit binary or parquet?
\pvted net
© & DingDongsingSong 8 montis, 2 weeks ap9
So what isthe answer to this question? Binary or Parquet? The files a Parquet. f youre simply copying afl, you just need to define the right
source type (ie. Parquet inthis instance. Why would you even consider Binary when te file isn* Binary type
puted 3 times
2 & kami 8 months, 3 weeks 290
Ive just tested iin Azure, created two Gen? storage accounts, used Binaty as source and destination, placed two parquet files in account one,
Created pipeline in ADF, added copy data activity and then define fst binary as source with wldeard path parquet) and the sink 3 binary, with
Inked service for account 2, selected Preservehierarchy it worked
puted 1 tines
© & Anshulsuryawanshi 9 montis 99
‘When using Binary dataset in copy activity, you can only copy from Binary datast to Binary dataset
upted times
© & Sandipau 10 mons, 8 weeks ago
this snouts be binary
puted tines
BM Verobon 11 montis, 1 week ago
‘The type property of the dataset must beset to Parquet
hitps//docs mirosof com/en-us/azure/data-actory/format-parquetiparquet-as-source
‘poted 2 nesYou have an Azure Data Lake Storage Gen? container that contains 100 TB of data
You need to ensure tha the data inthe containers available for read workloads in a secondary region ifn outage occurs inthe primary region.
‘The solution must minimize costs,
Which ype of data redundancy should you use?
‘A. geoedundant storage (GRS)
8. readsaccess georedundant storage (RA-GRS)
. zone-redundant storage (288)
© ocaly redundant storage (URS)
Correct Answer: 8
‘Geoedundant storage (wth GRS or G2RS) replicates your data to another physical location inthe secondary region to protect against regional
outages.
However that data is avaabe to be read only if he customer or Microsoft intiates a fallover from the primar to secondary region. When you
enable read access tothe secondary region, your data is avallable tobe read at al times, including ina situation where the primary region
becomes unavailable
Incorrect Answers
‘A While Georedundant storage (GRS) is cheaper than Read-Acoess Geo-Redundant Storage (RA-GRS), GRS does NOT inate automatic
tallover
©,0: Locally redundant storage (LRS) and Zoneedundant storage (ZRS) provides redundancy within a single region
Reference:
btpsifdacs. microsoft. com/en-us/azute/storage/common/storageredundancy
Community vote dtrbution
© & meet GETTATS)1 year, months aga
Bisright
Geo-redundant storage (with GRS or GZRS) replicates your data to another physical location inthe secondary region to protect against regional
cutages. However, that data is available to be read only if the customer or Microsof intites a failover fom the primary to secondary region.
‘When you enable read access tothe secondary region, your data is avaiable to be read a all times, including ina stuation where the primary
region becomes unavailable,
puted 76 times
2 & devader 10 moths, 4 weeks 350
‘looks corect answer, RA-GRS s always avalible because its auto failover Since this not asked inthe question but mere importantly the
question is about reducing cost which GRS.
‘voted 20 ines
© kenmexam 1 month 290
‘The question clearly says “savallable for read workloads in a secondary region’. Thisis only available when choosing RA-GRS." With GRS,
when a ciaster happens in the primary region the user hasta imitate 3 fever so thatthe secondary region becomes the primary region".
[Ate point you are reading from your secondary regian with GRS. Hence i believe the answers should be B
“https/ear microsoft. com/en-us/azure/storage/common/storage-redundancy#geo-redundant-storage
“https: /learn microsoft com/en-us/azure/storage/comman/storage-disater-recovery-guidance
ote Times
BB BRIO 9 months, 3 weeks 390
It should be A because of two reasons:
1. Minimize cost
2, When primaty is unavailable
Hence No need for RA.GRS.
ted tines
© & illyob0604 3 dys, 18 hours spo
Exactly. This is the point It clearly states" case of an outage! RA-GRS --> secondary region can be read also notin a case of outage
pated ines
© & Sashain San Francisco (GIES) 1 yo" 399
Imrmy opinion, | believe the and answers A and this why.
Inthe question they state *. available for read workloads in a secondary region IF AN OUTAGE OCCURES inthe primary.” Wel, answer B (RA-GRS)
states in Microsoft dacumentation that RA-GRS i for when". your datas avalable tobe read AT ALL TIMES, including in situation where theprimary region becomes unavailable”
‘To me, the nature ofthe qu
nis what isthe cheapest solution which allows for failover to read workload, when there isan outage. Answer (8)
‘Common sense would be A’ too because that is probably the most often reab-ife use case
upvted $8 ines
© & Sabadamal20r0atGmall 11 months, 1 week ago
It’s not about common sense rather about technology. With GRS, data remains available even if an entire data center becomes unavailable orf
‘theres a widespread regional failure. There would be a down time when a region becomes unavailable. Alternately, you could implement read
access geo-redundant storage (RA-GRS), which provides read-access tothe data in altemate locations.
puted nae
© & Rrko7 (Henna) 5 days, 19 hours ago
Read-access geo-redundant storage (RA-GRS) is correct answer,
pote T tines
PB OléSehool 2 weeks, s day 250
RA-GRS is always ON if selected no matter regional outage, GRS activates only in case of regional outage and thst wae the question, alsa it's
puted T ines
© & Mvia1029 1 month ago
‘With GRS or GZRS, the data in the secondary region int available fr read or write access unless there isa fllover tothe secondary region.
For read access tothe secondary region, configure your storage account to use
Read access geo redundant storage (RA GRS)
>Read access geo zone redundant storage (RA GZRS),
pte Times
2 & prosery 2 months 290
Bis for sure wrong Answer
Comect Answeris A
Because question ask minimum cost. which i A,
Furthermore, RA-GRS is always readable in 2nd region even if you don't have any availabilty issue in primary region. So Correct ansuer is A
puted 7s
© yyphnh 3 months, 1 week ago
GGRS does NOT intste automatic falover, So Bis correct
‘puted T ines
© kevint1210 3 months, 2 weeks ago
Should be A
‘puted 2 ines
2 bagustte 3 months, 3 weeks ago
‘is cheaper
pted 3 times
© & aurorafang 3 months 3 weeks 290
also vote for A after reading the comments
puted Times
BB uty 3 months, 3 weeks ago
read-access geo-redundant storage (RA-GRS) isthe answer asin the question it's mentioned, Only read access to be available
upvted Ttines
© & dainh225 «months, 1 week age
RS ws Ra GRS
The diference between GRS and RA GRS is faity simple, GAS ony allows to be ead inthe secondary zone int
primary o secondity while RA GRS allaws the option to read in the secondary whenever.
1 even of» filover from the
pte tines
2 & Jainh225 4 months, 1 week age
SRS vs a GRS
‘The diference between GRS and RA GAS is fairly simple, GRS only allows to be read in the secondary zone inthe even of 2 failover from the
primary to secondary while RA GRS allows the option to read in the secondary whenever.pte tines
2 & prathamesh1996 6 months, 2 weeks 290
‘Ais Cotrect for Minimize cost § When primary is unavailable,
pte tines
© & Andushi7 months, 1 week age
A because of costs aspect
‘pte nes
© muove 7 mnths, 3 weeks ago
Als correct because of cost, RA-GRS will cost $5,910.73, GRS wil cost 4586.12
pte 3 tines
BA Dicer months, 1 week 90
I you go to open a Azure storage account and choose GRS/ZRS/LRS/RA-GRS, there wll bea notice reminding you which options ae cheaper.
From cheapest to most expensive: LAS The (complete) DataCenter, not the region and not components inside a DataCenter.
So, what helps us inthis situation?
RS: "copies your data synchronously thre times within a single physical location in the primary region.” importants here the SINGLE PHYSICAL
LOCATION (meaning inside the same Data Center Sin our scenaia all copies woulkn’ work anymore)
=> Cis wrong.
ZRS:" copies your data synchronously across three Azure availability Zones inthe primary region” (meaning, in different Data Centers. ln our
Scenario this would meet the requirements)
=> Dis right
{GRS/GZAS: are lke LRS/ZRS but withthe Data Centers in diferent azure regions. Ths works too but is more expensive than ZS. So ZRS isthe right
hitps/docs microsoft com/en-us/azure/storage/common/storage-redundancy
pres 49 tines
8 & Detar 11 months ago
I agree.
Please give ths comment medal {ora cookie,
‘puted 3 nes
BM Orrens months, 2 weeks ago
Yes, well sad, that's the correct answer
ted I nee© Narasimhap 9 months, 3 weaks ago
Well explained!
pote ines
© & greentever (ETT) 1 month, 3 wesks 290
Microsoft recommends using ZRS in the primary region for Azure Data Lake Storage Gen2 workloads
pte Times
© A Deekshat234 3 months, 2 weeks ago
Dis correct
pote Times
© & RUKOT months, 1 weak ago
Dis corract ast talks abo,
pte Times
Gata center” means we cannot use the LRS (LOCAL)
© i olavrabs 6 months. 3 weeks 90
[D> Data is replicated synchronously
poted tines
© & Egocenttle 7 months, 2 weeks ago
Discorrect
puted 2simes
2 & ravi2931 7 montis, 3 weeks ago
it snoule be 0
‘puted Tinas
© & ravi2931 7 montis, 3 neeks ago
See ths explained leary
LAS isthe lowest-cost redundancy option and offers the least durability compared to other options. LRS protects your data against server rack
and drive falures. However ifa disaster such a fire or flooding occurs within te data center al replicas ofa storage account using LRS may be
lost or unrecoverable. To mitigate this risk, Microsoft recommends using zone-redundant storage (285), geo-redundant storage (GRS), oF geo-
zone-redindant storage (GZRS)
ted | nee
© & AS61208 8 mons 290
‘Answer shouldbe D, as LRS won't be helpfull incase of whole datacenter flr
pote Times
BA Andy91 8 months, 1 weak ego
Thisis the corect answer indeed
‘pted times
© & bhanuprasad9331 9 months, 1 week ago
‘Answer is LRS,
From microsoft docs:
RS replicates data ina single AZ. An AZ can contain one of more data centers. So, even fone dats center fils, dats can be accessed through
other datacenters inthe same AZ,
hitps//docs microsoft com/en-us/azure/avalabilty-zones/az-overvewtavaiabilty-20Nes
hitps/docs mirosof com/en-us/azure/storage/common/storage-redundancy#redundancyin-the-primary-region
puted times
© & PallaiPatel 10 montis, 1 weak 290
Dis correct
puted 3 mes
© & vimalnits 10 monts, 1 week 290
Comect answer's.
‘pte 2 times
© & Tinsaaaasa 10 months, week ago
LAS helps to replicate data in the same data center while ZS replicates data synchronously actoss three storage clusters in one regionpte T ines
© Shatheesh 10 months, 2 weoks ago
Dis the corect answer, In question its clealy mentioned if datacenter alt shouldbe available, LRS stores everything in sane data center so it's
not the correct answer, nest cheapest option fx ZRS.
pte T ines
1B & Jaws1990 11 months age
Mentions data centre Availability Zone) filre, not rack fale, so should be Zone Redundant Storage,
‘pvted 3 nas
BB Detar 11 months ago
note thatthe "lata centre fal
pvted 7 nasHoTsPor
You have a SQL pool in Azure Synapse
You plant load data from Azure Blob storage toa staging table, Approximately 1 milion rows of data willbe loaded daily. The tble will be
‘wuncated before each dally loa,
You need to create the staging table. The solution must minimize how long it takes to load the data to the staging table
How should you configure the table? To answer, select the appropriate options inthe answer area
NOTE: Each correct selection I worth one pont.
Hot aren
Answer Area
Distribution: v
[Hash
Replicated
[Round-robin
Indexing:
Partitioning:
Answer Area
Distribution: mA
Replicated
[Round-robin
comet Waexing =
[Clustered 7
|Clustered columnstore_|
[Heap
Partitioning: Vv
|None
Box: Hash
Hashstrbuted tables improve query psfrmanceonlarge fact bles, They can have vr ge numbers of ms and sil achieve high
perfomance
Incorrect Answers
Round-obin tables ae useful for improving loading speed.
Box 2; Clustered columnstore-
‘When creating partitions on clustered columnstore tables, itis important to consider how many rows belong to each partion For optimal
compression and performance of clustered columnstore tables, @ minimum of 1 milion cows per distribution an partition is needed.
Box: Date-
‘Table partons enable you to divide your data into smaller groups of data, In most cases, table partitions are created ona date column,
Pertti switching canbe used to quickly remove or replace a section ofa tableReference:
tpsifdocs.mierosoft.com/enus/azute/synapseanalyics/sql-data-warehouse/safdata warehouse tables partion
bttpsi/does.microsoftcom/en-us/azute/synapse-analytcs/sqtdate warehouse/sob-data warehouse tables distribute
© & av000 (EETaTa) yor, mnths 50
Round-Robin
Heap
None
puted 298 tines
BM anto69 10 months, 2 weeks a0
agree too
‘voted 3 nee
1B & Narasimhap 9 months, 5 weaks ago
Round Rebin
Heap
None.
No brainer fortis question
puted 7 nas
© & gssdéscoder 1 yea 1 month ago
Agree 100%,
Allin paragraphs under this: htps//docs microsoftcomyen-us/azure/synapse-analytics/sa/develop-tables-overview.
‘poted Sines
BB rtaz 11 month ago
Also agree 100%
‘pated 2imes
© A Deepshikhat228 4 months, 7 week ago
agree, Round Robin, Heap and None i the correct option
puted imas
© & laszek (HETTECTS] yo, 9 months ago
Round-robin - this iste simplest clstribution model, not great for querying but fast to process
Heap -no brainer when cresting staging tables
No partitions this isa staging table, why add effort to patton, when truncated daly?
‘pvted 40 tines
© berserkeap 1 yea: 1 month 290
Had doubts regarding why there is no need fora partion. While what you suggested is true wart it be better if there isa date partition to
truncate the table ?
pote Ives
BM andy.g 9 months, 3 weeks ago
‘Ther sno fiker on 3 truncate statement so na benefit in having » partion
‘ted ines
© & Vardhan_Brahmanapally 1 yea 292
‘Can you explain me why should we use heap?
‘voted T nes
8 & Detar 11 months ago
The term heap basicaly refers toa table without a clustered index. Adding a clustered index to a temp table makes absolutely no sense and
ita waste of compute resources fora table that would be entirely truncated daly
no clstered index = heap.
pote tines
© & saLDev0000 8 rons, 1 Week 90
Drtazis right in ation, when you populate an indexed table, you are also wrting tothe index, so this adds an adltional overhead in
the write process
poted 7 ines
© & peko7 (ETAT) days, rho ago
‘Ast only about stage table soit should be
Round-Robin|
Heap
None
pte | ines
© A rohitbinnant 1 wee, &43ys 350