Pentaho Data Integration
InstallationBusiness AnalyticsCToolsData CatalogData QualityLLMs
  • Overview
    • Pentaho Data Integration ..
  • Data Integration
    • Getting Started
      • Configuring PDI UI
      • KETTLE Variables
    • Concepts & Terminolgy
      • Hello World
      • Logging
      • Error Handling
    • Data Sources
      • Flat Files
        • Text
          • Text File Input
          • Text File Output
        • Excel
          • Excel Writer
        • XML
          • Read XML
        • JSON
          • Read JSON
      • Databases
        • CRUID
          • Database Connections
          • Create DB
          • Read DB
          • Update DB
          • Insert / Update DB
          • Delete DB
        • SCDs
          • SCDs
      • Object Stores
        • MinIO
      • SMB
      • Big Data
        • Hadoop
          • Apache Hadoop
    • Enrich Data
      • Merge
        • Merge Streams
        • Merge Rows (diff)
      • Joins
        • Cross Join
        • Merge Join
        • Database Join
        • XML Join
      • Lookups
        • Database Lookups
      • Scripting
        • Formula
        • Modified JavaScript Value
        • User Defined Java Class
    • Enterprise Solution
      • Jobs
        • Job - Hello World
        • Backward Chaining
        • Parallel
      • Parameters & Variables
        • Parameters
        • Variables
      • Scalability
        • Run Configurations
        • Partition
      • Monitoring & Scheduling
        • Monitoring & Scheduling
      • Logging
        • Logging
      • Dockmaker
        • BA & DI Servers
      • Metadata Injection
        • MDI
    • Plugins
      • Hierarchical Data Type
  • Use Cases
    • Streaming Data
      • MQTT
        • Mosquitto
        • HiveMQ
      • AMQP
        • RabbitMQ
      • Kafka
        • Kafka
    • Machine Learning
      • Prerequiste Tasks
      • AutoML
      • Credit Card
    • RESTful API
    • Jenkins
    • GenAI
  • Reference
    • Page 1
Powered by GitBook
On this page
  1. Data Integration
  2. Enterprise Solution

Parameters & Variables

PreviousParallelNextParameters

Last updated 12 months ago

Both variables and parameters in Pentaho Data Integration enable you to create a more dynamic and reusable pipeline.

Variables are often used within transformations and jobs for storing values, while parameters are mainly used in jobs to pass dynamic values to transformations.

Let's expand on that ..

Think of a parameter as a local variable .. They are reusable inputs that apply only to the specific transformation or job, they are defined in. When defining a parameter, you can assign it a default value to use or you can dynamically fetch it.

Variables are used to store values that can be used across multiple jobs and transformations. They can be defined at different levels of scope, such as global or local. So, for example, suppose you have multiple transformations that read data from different CSV files and write it to a database. You can define a variable for the database connection string and use it across all of these transformations.

Workshops

In Pentaho Data Integration (PDI), parameters are variables that allow you to make your ETL (Extract, Transform, Load) processes more dynamic and reusable. They can be used to pass values into your transformations or jobs, making it easier to customize and control the behavior of your data integration processes.

Variables can be used throughout Pentaho Data Integration, including in transformation steps and job entries. You define variables by setting them with the Set Variable step in a transformation or by setting them in the kettle.properties file in the directory.

Parameters
Variables