Your success in Microsoft 70-767 is our sole target and we develop all our 70-767 braindumps in a way that facilitates the attainment of this target. Not only is our 70-767 study material the best you can find, it is also the most detailed and the most updated. 70-767 Practice Exams for Microsoft MCSA 70-767 are written to the highest standards of technical accuracy.
Q18. You are designing a SQL Server Integration Services (SSIS) package that uses the Fuzzy Lookup transformation.
The reference data to be used in the transformation does not change.
You need to reuse the Fuzzy Lookup match index to increase performance and reduce maintenance.
What should you do?
A. Select the GenerateAndPersistNewIndex option in the Fuzzy Lookup Transformation Editor.
B. Select the GenerateNewIndex option in the Fuzzy Lookup Transformation Editor.
C. Select the DropExistingMatchlndex option in the Fuzzy Lookup Transformation Editor.
D. Execute the sp_FuzzyLookupTableMaintenanceUninstall stored procedure.
E. Execute the sp_FuzzyLookupTableMaintenanceInvoke stored procedure.
Answer: A
Explanation:
Reference: http://msdn.microsoft.com/en-us/library/ms137786.aspx
Q19. DRAG DROP
You are building a fact table in a data warehouse.
The table must have a columnstore index. The table cannot be partitioned. You need to design the fact table and load it with data.
Which three actions should you perform in sequence? (To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.)
Answer:
Q20. You are reviewing the design of a customer dimension table in an existing data warehouse hosted on SQL Azure.
The current dimension design does not allow the retention of historical changes to customer attributes such as Postcode.
You need to redesign the dimension to enable the full historical reporting of changes to multiple customer attributes including Postcode.
What should you do?
A. Add StartDate and EndDate columns to the customer dimension.
B. Add an IsCurrent column to the customer dimension.
C. Enable Snapshot Isolation on the data warehouse.
D. Add CurrentValue and PreviousValue columns to the customer dimension.
Answer: A
Q21. DRAG DROP
You administer a Microsoft SQL Server database. Service accounts for SQL Agents are configured to use a local user. A Microsoft SQL Server Integration Services (SSIS) job step has been created within a SQL Server Agent job. The SSIS package accesses a network share when exporting data from a SQL Server database.
When you execute the SQL Server Agent job, it fails due to a permissions failure on a share on a remote server. You need to ensure that the SQL Server Agent job can execute the SSIS package. Which four actions should you perform in sequence? (To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.)
Answer:
Explanation:
References:
http://msdn.microsoft.com/en-us/library/ms175834.aspx http://msdn.microsoft.com/en-us/library/ms189522.aspx
http://msdn.microsoft.com/en-us/library/ms190703.aspx http://msdn.microsoft.com/en-us/library/ms161950.aspx
Q22. You are designing a data warehouse for a fresh food distribution business that stores sales by individual product. It stores sales targets by product category. Products are classified into subcategories and categories.
Each product is included in only a single product subcategory, and each subcategory is included in only a single category.
The data warehouse will be a data source for an Analysis Services cube. The data warehouse contains two fact tables:
•factSales, used to record daily sales by product
•factProductTarget, used to record the monthly sales targets by product category
Reports must be developed against the warehouse that reports product sales by product, category and subcategory, and product sales targets.
You need to design the product dimension. The solution should use as few tables as possible while supporting all the requirements.
What should you do?
A. Create two product tables, dimProduct and dimProductCategory. Connect
factSales to dimProduct and factProductTarget to dimProductCategory with foreign key constraints. Direct the cube developer to use key granularity attributes.
B. Create one product table, dimProduct, which contains product detail, category, and subcategory columns. Connect factSales to dimProduct with a foreign key constraint. Direct the cube developer to use a non-key granularity attribute for factProductTarget.
C. Create three product tables, dimProduct, dimProductCategory, and dimProductSubcategory, and a fourth bridge table that joins products to their appropriate category and subcategory table records with foreign key constraints. Direct the cube developer to use key granularity attributes.
D. Create three product tables, dimProduct, dimProductCategory, and dimProductSubcategory. Connect factSales to all three product tables and connect factProductTarget to dimProductCategory with foreign key constraints. Direct the cube developer to use key granularity attributes.
Answer: B
Q23. You are using SQL Server Data Tools to develop a SQL Server Integration Services (SSIS) project.
The first package that you create in this project contains a package connection that accesses a flat file. Additional packages in the project must also access this file.
You need to define and reuse the flat file connection in all project packages. What should you do?
A. Convert the package Connection Manager in the first package to a project Connection Manager.
B. Copy the package Connection Manager and paste it into the second package.
C. Convert the project to the Package Deployment model.
D. Set the ProtectionLevel property of the package Connection Manager to DontSaveSensitive to reuse the flat file connection.
Answer: A
Q24. HOTSPOT
A SQL Server Integration Services (SSIS) package is designed to download data from a financial database hosted in SQL Azure.
The connection string to the financial database is defined as a project parameter named FinConStr. The parameter value must be stored securely and must be set explicitly every time the package is executed.
You need to configure the parameter to meet the requirements.
What should you do? (To answer, configure the appropriate option or options in the dialog box in the answer area.)
Answer:
Q25. You are editing a SQL Server Integration Services (SSIS) package that contains three Execute SQL tasks and no other tasks. The three Execute SQL tasks modify products in staging tables in preparation for a data warehouse load.
The package and all three Execute SQL product tasks have their TransactionOption property set to Supported.
You need to ensure that if any of the three Execute SQL product tasks fail, all three tasks will roll back their changes.
What should you do?
A. Change the TransactionOption property of the package to Required.
B. Change the TransactionOption property of all three Execute SQL product tasks to Required.
C. Move the three Execute SQL product tasks into a Foreach Loop container.
D. Move the three Execute SQL product tasks into a Sequence container.
Answer: A
Explanation:
References:
http://msdn.microsoft.com/en-us/library/ms137690.aspx http://msdn.microsoft.com/en-us/library/ms141144.aspx
Q26. You are creating a SQL Server Integration Services (SSIS) package to retrieve product data from two different sources. One source is hosted in a SQL Azure database. Each source contains products for different distributors.
Products for each distributor source must be combined for insertion into a single product table destination.
You need to select the appropriate data flow transformation to meet this requirement. Which transformation types should you use? (Each correct answer presents a complete
solution. Choose all that apply.)
A. Multicast
B. Merge Join
C. Term Extraction
D. union All
E. Merge
Answer: D,E
Explanation:
Reference: http://msdn.microsoft.com/en-us/library/ms141703.aspx
Reference: http://msdn.microsoft.com/en-us/library/ms141775.aspx
Reference: http://msdn.microsoft.com/en-us/library/ms141020.aspx
Reference: http://msdn.microsoft.com/en-us/library/ms141809.aspx
Reference: http://msdn.microsoft.com/en-us/library/ms137701.aspx
Q27. You are using a SQL Server Integration Services (SSIS) project that is stored in the SSIS catalog. An Environment has been defined in the SSIS catalog.
You need to add the Environment to the project. Which stored procedure should you use?
A. catalog.set_environment_reference_type
B. catalog.set_environment_property
C. catalog.create_environment_reference
D. catalog.create_environment
Answer: C
Q28. You are developing a project that contains multiple SQL Server Integration Services (SSIS) packages. The packages will be deployed to the SSIS catalog. One of the steps in each package accesses an FTP site to download sales transaction data.
You create project parameters to store the username and password that are used to access the FTP site.
You need to ensure that the username and password values are encrypted when they are deployed.
What should you do?
A. Set the Sensitive property of the parameters to True.
B. Set the ProtectionLevel property of the package to EncryptSensitiveWithUserKey.
C. Change the parameters to package parameters.
D. Change the project to the Legacy Deployment model.
Answer: A
Q29. You are deploying a new SQL Server Integration Services (SSIS) project to the test environment.
A package in the project uses a custom task component.
You need to ensure that the custom object is deployed on the test environment correctly. What should you do?
A. Run the package by using the dtexec /rep /conn command.
B. Create a reusable custom logging component.
C. Create an OnError event handler.
D. Use the gacutil command.
E. Use the dtutil /copy command.
F. Deploy the package to the Integration Services catalog by using dtutil and use SQL Server to store the configuration.
G. Run the package by using the dtexec /dumperror /conn command.
H. Use the Project Deployment Wizard.
I. Deploy the package by using an msi file.
J. Add a data tap on the output of a component in the package data flow.
K. Run the package by using the dtexecui.exe utility and the SQL Log provider.
Answer: D
Explanation:
Reference:
http://msdn.microsoft.com/en-us/library/ms403356.aspx
Q30. You are deploying a project to the SQL Server Integration Services (SSIS) catalog.
You need to ensure that the project is deployed to the production environment with the least amount of administrative effort.
What should you do?
A. Create an On Error event handler.
B. Add a data tap on the output of a component in the package data flow.
C. Use the gacutil command.
D. Create a reusable custom logging component.
E. Run the package by using the dtexecui.exe utility and the SQL Log provider.
F. Run the package by using the dtexec /rep /conn command.
G. Use the Integration Services Deployment Wizard.
H. Run the package by using the dtexec /dumperror /conn command.
I. Use the dtutil /copy command.
J. Deploy the package to the Integration Services catalog by using dtutil and use SQL Server to store the configuration.
K. Deploy the package by using an msi file.
Answer: G
Explanation:
Reference:
http://msdn.microsoft.com/en-us/library/hh479588.aspx http://msdn.microsoft.com/en-us/library/hh213290.aspx http://msdn.microsoft.com/en-us/library/hh213373.aspx
Q31. You are developing a data flow transformation to merge two data sources. One source contains product data and the other source contains data about the country in which the product was manufactured. Both data sources contain a two-character CountryCode column and both use SQL Server. Both data sources contain an ORDER BY clause to sort the data by the CountryCode column in ascending order.
You use a Merge Join transformation to join the data.
You need to ensure that the Merge Join transformation works correctly without additional transformations.
What should you do? (Each correct answer presents part of the solution. Choose all that apply.)
A. Change the ORDER BY clause on the product source to order by ProductName.
B. Change the Merge Join transformation to a Merge transformation.
C. Set the appropriate SortKeyPosition properties on the data sources.
D. Set the IsSorted property on both data sources.
Answer: C,D
Q32. DRAG DROP
You are building a SQL Server Integration Services (SSIS) package to load product data sourced from a SQL Azure database to a data warehouse. Before the product data is loaded, you create a batch record by using an Execute SQL task named Create Batch. After successfully loading the product data, you use another Execute SQL task named Set Batch Success to mark the batch as successful.
You need to create and execute an Execute SQL task to mark the batch as failed if either the Create Batch or Load Products task fails.
Which three steps should you perform in sequence? (To answer, move the appropriate
actions from the list of actions to the answer area and arrange them in the correct order.)
Answer:
Explanation:
References:
http://msdn.microsoft.com/en-us/library/ms141003.aspx http://msdn.microsoft.com/en-us/library/ms141261.aspx http://msdn.microsoft.com/en-us/library/ms141722.aspx http://msdn.microsoft.com/en-us/library/ms139895.aspx
Q33. DRAG DROP
You are validating whether a SQL Server Integration Services (SSIS) package named Master.dtsx in the SSIS catalog is executing correctly.
You need to display the number of rows in each buffer passed between each data flow component of the package.
Which three actions should you perform in sequence? (To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.)
Answer:
Explanation:
1. Run with verbose settings
2. Get Execution ID from .executions
3. Get stats from .execution_data_statistics ( rows_sent ) Ref: http://msdn.microsoft.com/en-us/library/hh230986.aspx
Q34. CORRECT TEXT
You are designing a package control flow. The package moves sales order data from a SQL Azure transactional database to an on-premise reporting database. The package will run several times a day, while new sales orders are being added to the transactional database.
The current design of the package control flow is shown in the answer area. (Click the Exhibit button.)
The Insert New Orders Data Flow task must meet the following requirements:
✑ Usage of the tempdb database should not be impacted.
✑ Concurrency should be maximized, while only reading committed transactions.
✑ If the task fails, only that task needs to be rolled back.
You need to configure the Insert New Orders Data Flow task to meet the requirements.
How should you configure the transaction properties? To answer, select the appropriate setting or settings in the answer area.
Answer:
IsolationLevel = ReadCommited. TransactionOption = Required