A customer has an existing PDI job that calls a transformation. They want to execute the transformation through Spark on their Hadoop cluster.
Which change must be made to satisfy this requirement?
You have a job that uses a Pentaho MapReduce entry to read four input files, and that outputs words and their counts to one output file.
How should you set the number of reducer tasks?
You are running a PDI Joband you need to identify when something fails for a specific job entry to notify the sysops team about the failure.
Which statement is correct in this scenario?
You have slow-running steps in a PDI transformation and you notice that it is taking a long time for subsequent steps to get data and begin processing.
Which action will help solve the problem?
A Big Data customer wants to run POI transformations on Spark on their production Hadoop cluster using Pentaho's Adaptive Execution Layer (AEL)
What are two steps for installing AEL? (Choose two.)
Choose 2 answers
A transformation is running in a production environment and you want to monitor it in real time.
Which tool should you use?
You want to manage a slowly changing dimension as type II.
Which step in PDI should you use to accomplish this task.
You execute a transformation locally in Spoon to read data from an Oracle database and you receive the Drive not found’ exception.
What are two ways to solve this problem? (Choose two.)
Choose 2 answers
You have a PDI job with an Oozie JobExecutor entry and you want this job entry to finish before the nextjob entry starts.
How do you accomplish this task?