

Medicaid Enterprise Data Warehouse
β - Featured Role | Apply direct with Data Freelance Hub
This role is a 12-month contract position based in Columbus, Ohio, requiring expertise in ETL/ELT processes, Big Data technologies, and data security. Key skills include Hadoop, Hive, Impala, and PySpark, with a focus on data ingestion and documentation.
π - Country
United States
π± - Currency
Unknown
-
π° - Day rate
-
ποΈ - Date discovered
August 19, 2025
π - Project duration
More than 6 months
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Columbus, OH
-
π§ - Skills detailed
#Security #Spark (Apache Spark) #Shell Scripting #Sqoop (Apache Sqoop) #Linux #Documentation #Data Integration #Cloudera #Data Security #EDW (Enterprise Data Warehouse) #Data Ingestion #Impala #StreamSets #Unix #Data Warehouse #Hadoop #Datasets #Scripting #Data Governance #Cloud #"ETL (Extract #Transform #Load)" #Big Data #HDFS (Hadoop Distributed File System) #Migration #Data Analysis #PySpark
Role description
LOCATION: Columbus, Ohio POSITION TYPE: 12 month contract POSITION TERMS: On site, 5 days a week Interview via Teams The consultant will be responsible for Enterprise Data Warehouse design, development, implementation, migration, maintenance and operation activities. Works closely with Data Governance and Analytics team. Will be one of the key technical resource for data warehouse projects for various Enterprise Data Warehouse projects and building critical Datamarts, data ingestion to Big Data platform for data analytics and exchange partners. Team Participation Participate in team activities, design discussions, stand-up meetings, and planning reviews β Data Analysis & Ingestion Perform data analysis, profiling, quality checks, and ingestion using Hadoop/Hive/Impala, PySpark, UNIX shell scripts β Code Quality Proven ability to write high-quality code β Data Handling Experience building datasets; familiarity with PHI and PII data β ETL/ELT Logic Expertise in implementing complex ETL/ELT processes β Reconciliation Develop and enforce strong reconciliation processes β Documentation Accountable for ETL/ELT design documentation β Data Security & Modeling Knowledge of Big Data, Hadoop, Hive, Impala, dimensional modeling, and data security β UNIX/Linux Basic knowledge of UNIX/Linux shell scripting β Big Data Projects Experience with Big Data, Hadoop in Data Warehousing or Data Integration projects Required 8 ETL/ELT with Cloudera Analysis, design, development, support, and enhancements of ETL/ELT using Cloudera Big Data technologies Required 7 Development (Big Data) Creating Sqoop scripts, PySpark programs, HDFS commands, file formats (Parquet, Avro, ORC), StreamSets pipelines, job scheduling Required 7 Hadoop/Hive/Impala Scripting Writing scripts to gather stats on tables post data loads Required 7
LOCATION: Columbus, Ohio POSITION TYPE: 12 month contract POSITION TERMS: On site, 5 days a week Interview via Teams The consultant will be responsible for Enterprise Data Warehouse design, development, implementation, migration, maintenance and operation activities. Works closely with Data Governance and Analytics team. Will be one of the key technical resource for data warehouse projects for various Enterprise Data Warehouse projects and building critical Datamarts, data ingestion to Big Data platform for data analytics and exchange partners. Team Participation Participate in team activities, design discussions, stand-up meetings, and planning reviews β Data Analysis & Ingestion Perform data analysis, profiling, quality checks, and ingestion using Hadoop/Hive/Impala, PySpark, UNIX shell scripts β Code Quality Proven ability to write high-quality code β Data Handling Experience building datasets; familiarity with PHI and PII data β ETL/ELT Logic Expertise in implementing complex ETL/ELT processes β Reconciliation Develop and enforce strong reconciliation processes β Documentation Accountable for ETL/ELT design documentation β Data Security & Modeling Knowledge of Big Data, Hadoop, Hive, Impala, dimensional modeling, and data security β UNIX/Linux Basic knowledge of UNIX/Linux shell scripting β Big Data Projects Experience with Big Data, Hadoop in Data Warehousing or Data Integration projects Required 8 ETL/ELT with Cloudera Analysis, design, development, support, and enhancements of ETL/ELT using Cloudera Big Data technologies Required 7 Development (Big Data) Creating Sqoop scripts, PySpark programs, HDFS commands, file formats (Parquet, Avro, ORC), StreamSets pipelines, job scheduling Required 7 Hadoop/Hive/Impala Scripting Writing scripts to gather stats on tables post data loads Required 7