

Database Engineer (Pentaho)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Database Engineer (Pentaho) with a contract length of over 6 months, offering $50.00 - $65.00 per hour. Key skills include ETL development using Pentaho, SQL databases, and data warehousing. Hybrid remote in East Pittsburgh, PA.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
520
-
ποΈ - Date discovered
August 13, 2025
π - Project duration
More than 6 months
-
ποΈ - Location type
Hybrid
-
π - Contract type
W2 Contractor
-
π - Security clearance
Unknown
-
π - Location detailed
East Pittsburgh, PA 15112
-
π§ - Skills detailed
#Storage #Database Administration #Data Pipeline #Cloud #GIT #Azure #Data Lake #Normalization #Microsoft Power BI #Hadoop #API (Application Programming Interface) #Data Integration #"ETL (Extract #Transform #Load)" #Scripting #Talend #BI (Business Intelligence) #Security #Data Security #MySQL #Data Warehouse #Python #SQL Server #Big Data #Logging #Complex Queries #Indexing #Oracle #AWS (Amazon Web Services) #Computer Science #Databases #Data Quality #Datasets #Spark (Apache Spark) #Groovy #Airflow #Scala #SQL (Structured Query Language) #Data Modeling #Batch #Base #Compliance #GCP (Google Cloud Platform) #Version Control #Dimensional Modelling #Data Governance #Informatica #Data Architecture #Tableau #Shell Scripting #PostgreSQL #Database Performance
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Strong experience in ETL development and data integration, particularly using Pentaho Data Integration (PDI). The ideal candidate will be responsible for designing, developing, and maintaining robust data pipelines and ETL processes to support enterprise-level data platforms. You will collaborate closely with data architects, analysts, and other stakeholders to ensure data quality, availability, and performance across the organization.
Key Responsibilities
Design, develop, and optimize ETL workflows using Pentaho Data Integration (Kettle).Maintain and improve existing ETL jobs and scheduling mechanisms for reliability and efficiency.Work closely with business and technical teams to understand data requirements and implement solutions.Integrate data from various sources (databases, flat files, APIs) into centralized data stores.Support data warehousing initiatives and ensure timely and accurate data delivery.Monitor and troubleshoot ETL jobs and database performance issues.Assist with database administration tasks such as indexing, partitioning, and backup strategies.Implement best practices for data security, quality, and governance.Document data workflows, schemas, and processes clearly for future maintenance and scalability.
Required QualificationsBachelorβs degree in Computer Science, Information Systems, or a related field.3+ years of experience in database engineering or ETL development.Strong proficiency in Pentaho Data Integration (PDI) β including job/transformation design, parameterization, and logging.Experience with SQL databases such as PostgreSQL, MySQL, SQL Server, or Oracle.Solid understanding of ETL principles, data modeling, and data warehousing concepts.Proficient in SQL and performance tuning for complex queries.Experience working with large datasets and batch processing environments.Familiarity with scheduling tools (e.g., cron, Airflow, or enterprise schedulers).Knowledge of version control systems (e.g., Git) and CI/CD workflowsExpertise in ETL tools specifically Pentaho Data Integrator β including transformations, jobs and scheduling.Strong hands-on experience with Oracle SQL and PL/SQL.Writing complex queries, procedures, packages, and performance tuning.Understanding of dimensional modelling and db normalizationExperience in building and maintaining data warehouses and martsFamiliar in shell scripting, git, data base administration, smtp.Strong problem solving and analytical skillsFamiliar with Tableau and other BI tools
Preferred QualificationsExperience with cloud-based data platforms (e.g., AWS, Azure, GCP).Familiarity with other ETL tools or frameworks (e.g., Talend, Informatica) is a plus.Knowledge of scripting languages such as Python, Shell, or Groovy.Exposure to big data technologies (e.g., Hadoop, Spark) or data lakes.Understanding of data governance, compliance, and security standards.
Job Types: Full-time, Contract
Pay: $50.00 - $65.00 per hour
Expected hours: No less than 40 per week
Benefits:
Dental insurance
Health insurance
Paid time off
Application Question(s):
Are you comfortable working on W2?
Are you legally authorized to work in the United States, US Citizen/Green Card Holder?
How many years of work experience do you have with Pentaho?
How many years of work experience do you have with API?
How many years of work experience do you have with Oracle?
How many years of work experience do you have with SQL?
How many years of work experience do you have with AWS and Azure and related tools?
How many years of work experience do you have with Tableau?
How many years of work experience do you have with Power BI?
How many years of work experience do you have with Shell Scripting?
How many years of work experience do you have with hadoop?
How many years of work experience do you have with Data lakes and storage?
Work Location: Hybrid remote in East Pittsburgh, PA 15112