Business-Objects DMDI301 Questions & Answers

Full Version: 33 Q&A






Business-Objects
DMDI301
BusinessObjects Data Integrator XI -�Level Two
Download Full Version : http://killexams.com/pass4sure/exam-detail/DMDI301
D. to_data(sales_date&' '& sales_time.'dd_mmmm-yyyy hh24:mi:ss')
Answer: A
QUESTION: 25
You have a production job that retrieves data from your oracle 10g operational source
system and loads the data into your data warehouse. The operations of the source system
are complaining that the data integrator load process is causing the system to perform
poorly. Which two actions can you take to reduce the impact data integrator jobs have
on the source system? (Choose two)
A. Implement a CDC data store for the source system t reduce the number of rows
extracted.
B. Increase the value of the "array _fetch_size" parameter on the source table.
C. Perform intensive operations such as "group by" and "joins" in a staging area instead
of on the source system.
D. Use "linked data stores" to connect the source and target data stores.
Answer: A, B
QUESTION: 26
Your data integrator environment interprets year values greater than 15 as 1915 instead
of 2015. you must ensure data integrator interprets any date from "00 to 90" as "2000 to
2090" without making direct modifications to the underlying data flow. Which method
should you use to accomplish this task?
A. Log into the designer and select tools l Options l data l General and modify the
"Century change year" to 90.
B. Open the server manger and select edit job server config and modify the "Century
change year to 90".
C. Open the web administration tool and select management l requisiteness edit
theproduction requsitury and modify the "Century change year to 90".
D. On the job server, open the windows l control panel l regional settings l Customize
data and modify the two digit year interpretation to 90.
E. Configure the source database to interpret the two digit dates appropriately.
Answer: A
QUESTION: 27
9
You load over 10,000.000 records from the "customer" source table into a staging area.
You need to remove the duplicate customer during the loading of the source table. You
do not nee tot record or audit the duplicates. Which two do-duplicating techniques will
ensure that best performance? (Choose two.)
A. Use a Query transform to order the incoming data set an use the previous_row-value
in the where clause to filter any duplicate row.
B. Use the Query transform to order the incoming data set. Then a table_comparison
transform with "input contains duplicates" and the "sorted input" options selected.
C. Use tha table_ comparison transform with the "input contains duplicates" and "cached
comparison table" selected.
D. Use the lookup_ext function. With the Pre_load_cache" option selected to test each
row for duplicates.
Answer: A, B
QUESTION: 28
You want to join the "sales" and "customer" tables. Both tables reside in different data
stores. The "sales" table contains approximately five million rows. The "customer" table
Contains approximately five thousand rows, the join occurs in memory. How would you
set the source table options to maximize the performance of the operation?
A. Set the sales table joins tank to 10 and the cache to "No" then set the customer table
join tank to 5 and cache to "yes".
B. Set the sales table joins tank to 10 and the cache to "yes" then set the customer table
join tank to 5 and cache to "yes".
C. Set the sales table joins tank to 5 and the cache to "Yes" then set the customer table
join tank to 10 and cache to "No".
D. Set the sales table joins tank to 5 and the cache to "No" then set the customer table
join tank to10 and cache to "No".
Answer: A
QUESTION: 29
Where can the XML_Pipeline transform be used within a data flow? (Choose two)
A. Immediately after an XML source file.
B. Immediately after an XML source message.
C. Immediately after a Query containing nested data.
D. Immediately after an XML template.
10
Answer: A, B
QUESTION: 30
You create a two stage process for transferring data from a source system to a target data
warehouse via a staging area. The job you create runs both processes in an overnight
schedule. The job fails at the point of transferring the data from the satging area to the
target data warehouse. During the work day you want to return the job without
impacting the source system and therefore want to just run the second stage of the
process to ransfer the data from the staging area to the data warehouse. How would you
design this job?
A. Create two data flows the first extracting the data from the source system the second
transferring the data to the target data warehouse.
B. Create one data flow which extracts the data form the source system and uses a
data_transfer transform to stage the data in the staging area before then continuing to
transfer the data to the target data warehouse.
C. Create two data flows the first extracting the data from the source system and uses a
data_tranfer transform to write the data to the staging area. The second data flow
extracts the data from the staging area and transfers it to the target data warehouse.
D. Create one data flow which extracts from the source system and populates both the
staging area and the target data warehouse.
Answer: A
QUESTION: 31
Which two data integrator objects/operations support load balancing in a server Group
based architecture? (Choose two.)
A. Job
B. Lookup_ext
C. Script
D. While loop
Answer: A, B
QUESTION: 32
You have a data flow the read multiple XML files form a directory by specifying
wildcard in the file name. which method can use to link the XML file name to the
records being read?
11
A. Select "include file name column" in the XML source file.
B. Use the function get_xml file name in the query mapping
C. Use the column "XML_fileNAME" listed at the top of the XML file structure.
D. Use the variable$ current_XML_file in
the query mapping
Answer: A
QUESTION: 33
You are trying to improve the performance of a simple data flow that loads data from a
source table into a staging area and only applies some simple remapping using a Query
transform. The source database is located on the wan. The network administrator has
told you that you can improve performance if you reduce the number or round trips that
occur between the data integrator job server and the source database. What can you do in
your data flow to achieve this?
A. increase the array reach size parameter in the source table editor
B. Increase the commit size in the target table editor.
C. Increase the commit size in the source table editor.
D. Replace the source table with the SQL transform.
Answer: A

12
For More exams visit http://killexams.com





Kill your exam at First Attempt....Guaranteed!

User: Nur*****

I needed an e-book that could provide questions and answers for the DMDI301 exam, and I found exactly what I required on Killexams.com. The platform was responsible for my success, and I am grateful to them for providing such a tremendous resource. I had attempted the DMDI301 exam for three consecutive years, but could not reach the passing marks. However, with Killexams.com, I finally understood my knowledge gaps and passed the exam with flying colors.
User: Luciano*****

If you want to ace your online dmdi301 exams, I have an outstanding and easy way to do so, and that is by using Killexams.com and their dmdi301 exam sample papers. These papers are probably a true representation of the final test for the dmdi301 exam. I scored 95% in the very last exam, and I can say that Killexams.com is a product for those who continuously want to improve themselves and want to perform a little extra every day. The dmdi301 trial test can enhance your self-confidence.
User: Sarah*****

I highly recommend killexams.com question bank to anyone planning to take these tests. Congratulations on a well-executed job. Thanks to their materials, I was able to pass my DMDI301 exams.
User: Zoey*****

I highly recommend killexams.com to anyone preparing for the DMDI301 certification exam. This platform offers excellent study materials that helped me pass the exam with confidence. Their practice tests and study guides provided top-notch preparation, boosting my knowledge and ensuring my success.
User: Shanina*****

Thanks to killexams.com, I was able to pass the DMDI301 exam with a score of 89%. Their exam arrangement was spot on and helped me tremendously.

Features of iPass4sure DMDI301 Exam

  • Files: PDF / Test Engine
  • Premium Access
  • Online Test Engine
  • Instant download Access
  • Comprehensive Q&A
  • Success Rate
  • Real Questions
  • Updated Regularly
  • Portable Files
  • Unlimited Download
  • 100% Secured
  • Confidentiality: 100%
  • Success Guarantee: 100%
  • Any Hidden Cost: $0.00
  • Auto Recharge: No
  • Updates Intimation: by Email
  • Technical Support: Free
  • PDF Compatibility: Windows, Android, iOS, Linux
  • Test Engine Compatibility: Mac / Windows / Android / iOS / Linux

Premium PDF with 33 Q&A

Get Full Version

All Business-Objects Exams

Business-Objects Exams

Certification and Entry Test Exams

Complete exam list