更新时间: 试题数量: 购买人数: 提供作者:

有效期: 个月

章节介绍: 共有个章节

收藏
搜索
题库预览
You have a Fabric tenant that contains a lakehouse named Lakehouse1.

您有一个名为Lakehousel的Fabric租户,其中包含一个湖仓。

Readings from 100 loT devices are appended to a Delta table in Lakehouse1. Each set of readings isapproximately 25 KB. Approximately 10 GB of data is received daily.

来自100个物联网设备的读数被追加到Lakehouse1中的Delta表中。每组读数大约为25KB。

每天接收约10GB的数据。

All the table and SparkSession settings are set to the default.

所有表和SparkSession设置均被重置为默认值。

You discover that queries are slow to execute. In addition, the lakehouse storage contains data and logfiles that are no longer used.

您发现查询执行速度缓慢。此外,湖仓存储中包含不再使用的数据和日志文件。

You need to remove the files that are no longer used and combine smallfiles into larger files with a targetsize of 1 GB per file.

你需要删除不再使用的文件,并将小文件合并为单个目标大小为1GB的大文件。

What should you do? To answer, drag the appropriate actions to the correct requirements. Each actionmay be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.

你该怎么做呢?要回答这个问题,请将相应的操作拖放到对应的要求栏中。每个操作可能仅使用一次多次使用或完全不用。你可能需要拖动窗格间的分割栏,或者滚动查看内容。

NOTE: Each correct selection is worth one point.

注意:每项正确选择得一分。

(含图)

Contoso, Ltd. is a US-based health supplements company. Contoso has two divisions named Sales andResearch. The Sales division contains two departments named Online Sales and Retail Sales. TheResearch division assigns internally developed product lines to individual teams of researchers andanalysts.

Contoso,Ltd.是一家总部位于美国的健康保健品公司。Contoso设有两个部门,分别名为销售部和研究部。销售部下设在线销售部和零售销售部两个部门。研究部则将内部开发的产品线分配给各个由研究人员和分析人员组成的团队。

Existing Environment 现有环境

Identity Environment 身份环境

Contoso has a Microsoft Entra tenant named contoso.com. The tenant contains two groups namedResearchReviewersGroup1 and ResearchReviewersGroup2.

Contoso拥有一个名为conttoso.com的MicrosoftEntra租户。该租户包含两个组,分别为ResearchReviewersGroup1ResearchReviewersGroup2.

Data Environment 数据环境

Contoso has the following data environment:

Contoso拥有以下数据环境:

- The Sales division uses a Microsoft Power BI Premium capacity.

- 销售部门使用Microsoft Power BI Premium容量。

- The semantic model of the Online Sales department includes a fact table named Orders that usesImport mode. In the system of origin, the OrderlD value represents the sequence in which orders arecreated.

- 在线销售部门的语义模型包含一个名为“订单”的事实表,该表采用导入模式。在源系统中,OrderlD值表示订单创建时的顺序。

-The Research department uses an on-premises, third-party data warehousing product.

- 研究部门使用一个本地部署的第三方数据仓库产品。

- Fabric is enabled for contoso.com.An Azure Data Lake Storage Gen2 storage account named storage1 contains Research division datafor a product line named Productline1. The data is in the delta format.

- Fabric 已为contoso.com启用。一个名为storagel的Azure Data Lake Storage Gen2存储账户包含研究部门的产品线Productlinel的数据。数据采用delta格式。

- A Data Lake Storage Gen2 storage account named storage2 contains Research division data for aproduct line named Productline2. The data is in the CSV format.

- 一个名为storage2的数据湖存储Gen2存储账户包含研究部门的产品线Productline2数据。数据为CSV格式。

Requirements 需求

Planned Changes 计划变更

Contoso plans to make the following changes:

Contoso计划进行以下更改:

Enable support for Fabric in the Power BI Premium capacity used by the Sales division.

在销售部门使用的Power BI Premium容量中启用对Fabric的支持。

Make all the data for the Sales division and the Research division available in Fabric.

在Fabric中使销售部门和研究部门的所有数据可用。

For the Research division, create two Fabric workspaces named Productline 1ws and Productline 2ws.

为研究部门创建两个名为Productline 1ws和Productline 2ws的Fabric工作区。

In Productline 1ws, create a lakehouse named Lakehouse1.

在Productline 1ws中,创建一个名为Lakehousel的Lakehouse。

In Lakehouse1, create a shortcut to storage1 named ResearchProduct.

在Lakehousel中,创建一个名为ResearchProduct的快捷方式指向storage1。

Data Analytics Requirements 数据分析需求

Contoso identifies the following data analytics requirements:

Contoso确定了以下数据分析师需求:

- All the workspaces for the Sales division and the Research division must support all Fabricexperiences.

- 销售部门和研究部门的所有工作区必须支持所有Fabric体验。

- The Research division workspaces must use a dedicated, on-demand capacity that has per-minute billing.

- 研究部门的工作区必须使用专用的按需容量,且支持按分钟计费。

- The Research division workspaces must be grouped together logically to support OneLake data hubfiltering based on the department name.

- 研究部门的工作区必须逻辑分组,以支持基于部门名称的OneLake数据枢纽筛选。

- For the Research division workspaces, the members of ResearchReviewersGroup1 must be able toread lakehouse and warehouse data and shortcuts by using SQL endpoints.

- 对于研究部门的工作区,ResearchReviewersGroup1的成员必须能够通过sQL端点读取湖仓和仓库数据,并使用快捷方式。

- For the Research division workspaces, the members of ResearchReviewersGroup2 must be able toread lakehouse data by using Lakehouse explorer.

- 对于研究部门的工作区,ResearchReviewersGroup2的成员必须能够通过Lakehouse浏览器读取湖仓数据。

- All the semantic models and reports for the Research division must use version control that supportsbranching.

- 研究部门的所有语义模型和报告必须使用支持分支的版本控制。

Data Preparation Requirements 数据准备要求

Contoso identifies the following data preparation requirements:

Contoso确定了以下数据准备需求:

- The Research division data for Productline2 must be retrieved from Lakehouse1 by using Fabricnotebooks.

- 研究部门的产品线2数据必须通过Fabric笔记本从Lakehouse1获取。

- All the Research division data in the lakehouses must be presented as managed tables in Lakehouseexplorer.

- 湖仓中的所有研究部门数据必须在湖仓浏览器中以管理表形式呈现。

Semantic Model Requirements 语义模型要求

Contoso identifies the following requirements for implementing and managing semantic models:

Contoso为实现和管理语义模型识别了以下需求:

- The number of rows added to the Orders table during refreshes must be minimized.

- 在刷新期间添加到Orderstable表的行数必须最小化。

- The semantic models in the Research division workspaces must use Direct Lake mode.

- 研究部门工作区中的语义模型必须使用DirectLake模式。

General Requirements 通用要求

Contoso identifies the following high-level requirements that must be considered for all solutions:

Contoso确定了以下所有解决方案必须考虑的高层级需求:

- Follow the principle of least privilege when applicable.

- 在适用情况下遵循最小权限原则。

- Minimize implementation and maintenance effort when possible.

- 在可能的情况下,尽量减少实施和维护工作量。

QUESTION 1

You need to recommend which type of Fabric capacity SKU meets the data analytics requirements for theResearch division.

你需要推荐哪种类型的Fabric容量SKU满足研究部门的数据分析需求。

What should you recommend?

你应该推荐什么?

A. A

B. EM

C. P

D. F

QUESTION 2

Which workspace role assignments should you recommend for ResearchReviewersGroup1 andResearchReviewersGroup2? To answer, select the appropriate options in the answer area.

您应为ResearchReviewersGroup1和ResearchReviewersGroup2推荐哪些工作区角色分配?请在答题区域选择合适的选项。

NOTE: Each correct selection is worth one point.

注意:每项正确选择得一分。

(含图)

QUESTION 3

You need to ensure that Contoso can use version control to meet the data analytics requirements and thegeneral requirements.

你需要确保Contoso能够使用版本控制来满足数据分析要求和一般要求。

What should you do?

你应该怎么做?

A. Store all the semantic models and reports in Data Lake Gen2 storage.

A. 将所有语义模型和报告存储在Data Lake Gen2存储中。

B.Modify the settings of the Research workspaces to use a GitHub repository.

B. 修改研究工作区的设置以使用GitHub仓库。

C.Modify the settings of the Research division workspaces to use an Azure Repos repository.

C. 修改研究部门工作区的设置,使用Azure Repos仓库。

D. Store all the semantic models and reports in Microsoft OneDrive.

D. 将所有语义模型和报告存储在Microsoft OneDrive中。

QUESTION 4

You need to recommend a solution to group the Research division workspaces.

你需要推荐一个解决方案来分组研究部门的工作区。

What should you include in the recommendation? To answer, select the appropriate options in the answer area

建议中应包含哪些内容?请在答题区域选择合适的选项。

NOTE: Each correct selection is worth one point.

注意:每项正确选择得一分。

(含图)

Litware, Inc. is a manufacturing company that has offices throughout North America. The analytics team atLitware contains data engineers, analytics engineers, data analysts, and data scientists.

Litware公司是一家在北美各地设有办公室的制造公司。Litware的分析团队包括数据工程师、分析工程师、数据分析员和数据科学家。

Existing Environment 现有环境

Fabric Environment Fabric环境

Litware has been using a Microsoft Power BI tenant for three years. Litware has NOT enabled any Fabriccapacities and features.

Litware已使用Microsoft Power BI租户三年。Litware尚未启用任何Fabric功能和特性。

Available Data 可用数据

Litware has data that must be analyzed as shown in the following table.

Litware有数据需要分析,如下表所示。

(含图)

The Product data contains a single table and the following columns.

产品数据包含一个单一的表格和以下列。

The customer satisfaction data contains the following tables:

客户满意度数据包含以下表:

- Survey 调查表

- Question 问题

- Response 响应

For each survey submitted, the following occurs:

对于每份提交的调查,会发生以下情况:

- One row is added to the Survey table.

- 一条记录被添加到调查表中。

- One row is added to the Response table for each question in the survey.

- 每个调查问题在响应表中增加一行。

The Question table contains the text of each survey question. The third question in each survey responseis an overall satisfaction score. Customers can submit a survey after each purchase.

问题表包含每个调查问题的文本。每次调查回复中的第三个问题是整体满意度评分。客户可以在每次购买后提交调查问卷。

User Problems 用户问题

The analytics team has large volumes of data, some of which is semi-structured. The team wants to useFabric to create a new data store.

分析团队拥有大量数据,其中部分为半结构化数据。团队希望使用Fabric创建一个新的数据存储系统。

Product data is often classified into three pricing groups: high, medium, and low. This logic is implementedin several databases and semantic models, but the logic does NOT always match across implementations

产品数据通常被划分为三个定价类别:高、中、低。这一逻辑被应用于多个数据库和语义模型中,但不同实现方式下的逻辑并不总是完全一致的。

Requirements 需求

Planned Changes 计划变更

Litware plans to enable Fabric features in the existing tenant. The analytics team will create a new datastore as a proof of concept (PoC). The remaining Litware users will only get access to the Fabric featuresonce the PoC is complete. The PoC will be completed by using a Fabric trial capacity.

Litware计划为现有租户启用Fabric功能。分析团队将创建一个新的数据存储作为概念验证(PoC)之用。其余的Litware 用户只有在PoC 完成之后才能访问 Fabric 功能。PoC 的完成将通过利用 Fabric 的试用容量来实现。

The following three workspaces will be created:

将创建以下三个工作区:

- AnalyticsPOC: Will contain the data store, semantic models, reports pipelines, dataflow, and notebooksused to populate the data store.

- AnalyticsPOC:将包含用于填充数据存储的数据存储、语义模型、报告管道、数据流和笔记本。

- DataEngPOC: Will contain all the pipelines, dataflows, and notebooks used to populate OneLake.

- DataEiPOC:将包含用于填充OneLake的所有管道、数据流和笔记本。

- DataSciPOC: Will contain all the notebooks and reports created by the data scientists

- DataSciPOC:将包含数据科学家创建的所有笔记本和报告

The following will be created in the AnalyticsPOC workspace:

以下内容将在AnalyticsPOC工作区中创建:

- A data store (type to be decided)

- 一个数据存储(类型待定)

- A custom semantic model

- 一个自定义语义模型

- A default semantic model

- 一个默认的语义模型

- Interactive reports

- 交互式报告

The data engineers will create data pipelines to load data to OneLake either hourly or daily depending onthe data source. The analytics engineers will create processes to ingest, transform, and load the data tothe data store in the AnalyticsPOC workspace daily. Whenever possible, the data engineers will use low-code tools for data ingestion. The choice of which data cleansing and transformation tools to use will be atthe data engineers' discretion.

数据工程师将创建数据管道,以便根据数据源的不同,以小时或日频度将数据加载至OneLake中分析工程师则将创建相关流程,用于每日将数据导入、转换并加载至Analytics POC工作区中的数据存储库中。在条件允许的情况下,数据工程师将使用低代码工具进行数据导入。选择使用何种数据清洗与转换工具则完全由数据工程师自行决定。

Allthe semantic models and reports in the Analytics POC workspace will use the data store as the soledata source.

AnalyticsPOC工作区中的所有语义模型和报告都将使用数据存储作为唯一的数据源。

Technical Requirements 技术要求

The data store must support the following:

数据存储必须支持以下功能:

- Read access by using T-SQL or Python

- 使用T-SQL或Python进行读取访问

- Semi-structured and unstructured data

- 半结构化和非结构化数据

- Row-level security (RLS) for users executing T-SQL queries

- 执行T-SQL查询的用户支持行级安全(RLS)

Files loaded by the data engineers to OneLake will be stored in the Parquet format and will meet DeltaLake specifications.

数据工程师加载到OneLake的文件将以Parquet格式存储,并符合Delta Lake规范。

Data will be loaded without transformation in one area of the AnalyticsPOC data store. The data will thenbe cleansed, merged, and transformed into a dimensional model.

数据将在AnalyticsPOC数据存储的一个区域中未经转换直接加载。随后,数据将被清洗、合并,并转化为维度模型。

The data load process must ensure that the raw and cleansed data is updated completely beforepopulating the dimensional model.

数据加载过程必须确保在填充维度模型之前,原始和清洗后的数据已完全更新。

The dimensional model must contain a date dimension. There is no existing data source for the datedimension. The Litware fiscal year matches the calendar year. The date dimension must always containdates from 2010 through the end of the current year.

该维度模型必须包含一个日期维度。目前尚无现成的日期维度数据源。Litware的财政年度与日历年度相吻合。日期维度必须始终包含从2010年至当前年结束期间的日期。

The product pricing group logic must be maintained by the analytics engineers in a single location. Thepricing group data must be made available in the data store for T-SQL queries and in the default semantic model. The following logic must be used:

产品定价组的逻辑必须由分析工程师在一个特定位置进行维护。定价组数据必须被存储在数据仓库中以供T-SQL查询使用,并纳入默认语义模型中。以下逻辑必须得到应用:

- List prices that are less than or equal to 50 are in the low pricing group.

- 价格小于或等于50的列表属于低价组。

- List prices that are greater than 50 and less than or equal to 1,000 are in the medium pricing group.

- 价格大于50且小于或等于1,000的商品属于中等定价组。

- List prices that are greater than 1,000 are in the high pricing group.

- 超过1,000的列表价格属于高价组。

Security Requirements 安全要求

Only Fabric administrators and the analytics team must be able to see the Fabric items created as part ofthe PoC.

只有Fabric管理员和分析团队才能查看作为PoC一部分创建的Fabric项目。

Litware identifies the following security requirements for the Fabric items in the AnalyticsPOC workspace:

Litware为Analytics POC工作区中的Fabric项目识别了以下安全要求:

- Fabric administrators will be the workspace administrators.

- Fabric管理员将是工作区管理员。

- The data engineers must be able to read from and write to the data store. No access must be grantedto datasets or reports.

- 数据工程师必须能够读取和写入数据存储。不得授予对数据集或报告的访问权限。

- The analytics engineers must be able to read from, write to, and create schemas in the data store. Theyalso must be able to create and share semantic models with the data analysts and view and modify allreports in the workspace.

- 数据分析工程师必须具备从数据存储中读取、向其中写入以及创建数据模式的能力。他们还必须能够与数据分析师共同创建并分享语义模型,并能查看及修改工作区内的所有报告。

- The data scientists must be able to read from the data store, but not write to it. They will access thedata by using a Spark notebook。

- 数据科学家必须能够从数据存储中读取,但不能写入。他们将通过使用Spark笔记本访问数据。

- The data analysts must have read access to only the dimensional model objects in the data store. Theyalso must have access to create Power BI reports by using the semantic models created by theanalytics engineers.

- 数据分析师必须仅对数据存储中的维度模型对象拥有读权限。他们还必须能够使用分析工程师创建的语义模型来创建Power B报告

- The date dimension must be available to all users of the data store.

- 日期维度必须对数据存储的所有用户可用。

- The principle of least privilege must be followed.

- 必须遵循最小权限原则。

Both the default and custom semantic models must include only tables or views from the dimensionalmodel in the data store. Litware already has the following Microsoft Entra security groups:

默认和自定义语义模型都必须仅包含数据存储中维度模型中的表或视图。Litware 已经拥有以下MicrosoftEntra安全组:

- FabricAdmins: Fabric administrators

- FabricAdmins:Fabric管理员

- AnalyticsTeam: All the members of the analytics team

- 分析团队:所有分析团队成员

- DataAnalysts: The data analysts on the analytics team

- 数据分析师:分析团队中的数据分析师

- DataScientists: The data scientists on the analytics team

- 数据科学家:分析团队中的数据科学家

- DataEngineers: The data engineers on the analytics team

- 数据工程师:分析团队中的数据工程师

- AnalyticsEngineers: The analytics engineers on the analytics team

- 分析工程师:分析团队中的分析工程师

Report Requirements 报告要求

The data analysts must create a customer satisfaction report that meets the following requirements:

数据分析师必须创建一份符合以下要求的客户满意度报告:

- Enables a user to select a product to filter customer survey responses to only those who havepurchased that product.

- 允许用户选择产品,以筛选仅针对已购买该产品的客户的调查回复。

- Displays the average overall satisfaction score of all the surveys submitted during the last 12 monthsup to a selected date.

- 显示截至选定日期的过去12个月内提交的所有调查问卷的平均整体满意度得分。

- Shows data as soon as the data is updated in the data store.

- 数据存储更新后立即显示数据。

- Ensures that the report and the semantic model only contain data from the current and previous year.

- 确保报告和语义模型仅包含当前和上一年的数据。

- Ensures that the report respects any table-level security specified in the source data store.

- 确保报告遵守源数据存储中指定的任何表级安全策略。

- Minimizes the execution time of report queries.

- 最小化报告查询的执行时间

Question 1

You need to assign permissions for the data store in the AnalyticsPOC workspace. The solution must meetthe security requirements.

您需要为AnalyticsPOC工作区中的数据存储分配权限。解决方案必须满足安全要求。

Which additional permissions should you assign when you share the data store? To answer, select theappropriate options in the answer area.

在共享数据存储时,应分配哪些额外权限?请在答案区域中选择适当的选项作答。

NOTE: Each correct selection is worth one point.

注意:每项正确选择得一分。

(含图)

1