Pentaho Data Integration
InstallationBusiness AnalyticsCToolsData CatalogData QualityLLMs
  • Overview
    • Pentaho Data Integration ..
  • Data Integration
    • Getting Started
      • Configuring PDI UI
      • KETTLE Variables
    • Concepts & Terminolgy
      • Hello World
      • Logging
      • Error Handling
    • Data Sources
      • Flat Files
        • Text
          • Text File Input
          • Text File Output
        • Excel
          • Excel Writer
        • XML
          • Read XML
        • JSON
          • Read JSON
      • Databases
        • CRUID
          • Database Connections
          • Create DB
          • Read DB
          • Update DB
          • Insert / Update DB
          • Delete DB
        • SCDs
          • SCDs
      • Object Stores
        • MinIO
      • SMB
      • Big Data
        • Hadoop
          • Apache Hadoop
    • Enrich Data
      • Merge
        • Merge Streams
        • Merge Rows (diff)
      • Joins
        • Cross Join
        • Merge Join
        • Database Join
        • XML Join
      • Lookups
        • Database Lookups
      • Scripting
        • Formula
        • Modified JavaScript Value
        • User Defined Java Class
    • Enterprise Solution
      • Jobs
        • Job - Hello World
        • Backward Chaining
        • Parallel
      • Parameters & Variables
        • Parameters
        • Variables
      • Scalability
        • Run Configurations
        • Partition
      • Monitoring & Scheduling
        • Monitoring & Scheduling
      • Logging
        • Logging
      • Dockmaker
        • BA & DI Servers
      • Metadata Injection
        • MDI
    • Plugins
      • Hierarchical Data Type
  • Use Cases
    • Streaming Data
      • MQTT
        • Mosquitto
        • HiveMQ
      • AMQP
        • RabbitMQ
      • Kafka
        • Kafka
    • Machine Learning
      • Prerequiste Tasks
      • AutoML
      • Credit Card
    • RESTful API
    • Jenkins
    • GenAI
  • Reference
    • Page 1
Powered by GitBook
On this page
  1. Data Integration
  2. Enterprise Solution
  3. Jobs

Job - Hello World

What's a Pentaho Job ..

PreviousJobsNextBackward Chaining

Last updated 4 months ago

In this guided demonstration, we will create a Job.

The concepts learnt, help to build the foundation necessary for creating any Job.

In this guided demonstration, you will:

  • Learn to create a new Job.

  • Add entries and configure hops.

START defines the starting point for job execution. Every job must have one (and only one) Start. Unconditional job hops only are available from a Start job entry. The start job entry settings contain basic scheduling functionality; however, scheduling is not persistent and is only available while the device is running.

The Data Integration Server provides a more robust option for scheduling execution of jobs and transformations and is the preferred alternative to scheduling using the Start step. If you want the job to run like a daemon process, however, enable Repeat in the job settings dialog box.

Note: The basic scheduling functionality and the repeat option are only functional within the main job and not within a sub job.

To create a newJob:

  1. In Spoon, click File > New > Job.

  2. Drag the ‘START’ job entry onto the canvas.

The Transformation job entry is used to execute a previously defined transformation. For ease of use, it is also possible to create a new transformation within the dialog, pressing the New Transformation button.

  1. Drag the ‘Transformation’ job entry onto the canvas.

  2. Double-click on the step, and configure the following properties:

This step clears any error state encountered in a job and forces it to a success state.

  1. Drag the ‘Success’ job entry onto the canvas.

  1. Click the Run button in the Canvas Toolbar.

  2. Click on the Job Metrics tab.

The Job Entries are executed sequentially.

Pentaho Job
Configure Transformation Job Entry
Job Metrics
Metrics