Richmond, VA- DATA ENGINEER

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer in Richmond, VA, as a 1099 Independent Contractor for a 6-month contract at a competitive pay rate. Key skills include Python, AWS, and PySpark, with a focus on ETL/ELT pipelines and data modeling.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
-
πŸ—“οΈ - Date discovered
September 24, 2025
πŸ•’ - Project duration
Unknown
-
🏝️ - Location type
Unknown
-
πŸ“„ - Contract type
1099 Contractor
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
Richmond, VA
-
🧠 - Skills detailed
#Programming #Airflow #Data Storage #"ETL (Extract #Transform #Load)" #Data Manipulation #Cloud #Big Data #Data Lake #Docker #Data Pipeline #Redshift #Kafka (Apache Kafka) #Python #Compliance #GIT #Computer Science #SQL (Structured Query Language) #Data Engineering #AWS Glue #PySpark #Automation #Agile #Data Quality #Data Science #S3 (Amazon Simple Storage Service) #Version Control #Data Governance #Spark (Apache Spark) #Athena #Data Modeling #Security #Datasets #Microsoft Power BI #Databases #Scala #NoSQL #Data Processing #Kubernetes #Lambda (AWS Lambda) #AWS (Amazon Web Services) #Tableau #BI (Business Intelligence) #Storage
Role description
We are seeking a skilled and detail-oriented Data Engineer with strong expertise in Python, AWS, and PySpark to join our team as a 1099 Independent Contractor. The ideal candidate will be responsible for designing, building, and maintaining scalable data pipelines, ensuring data quality, and enabling advanced analytics and reporting across the organization. Key Responsibilities β€’ Design, develop, and optimize ETL/ELT data pipelines using Python, PySpark, and AWS services. β€’ Ingest, transform, and process large-scale datasets from various structured and unstructured sources. β€’ Work with cloud-native tools (AWS Glue, Lambda, EMR, S3, Redshift, Athena, etc.) to manage data storage, transformation, and access. β€’ Implement and maintain data models, schemas, and data lakes/warehouses. β€’ Ensure data quality, reliability, and availability across all stages of the pipeline. β€’ Collaborate with data scientists, analysts, and business teams to understand data requirements and deliver solutions. β€’ Monitor pipeline performance, troubleshoot issues, and implement best practices for optimization and security. β€’ Document processes, workflows, and data flows to support long-term scalability and team knowledge-sharing. Required Skills & Qualifications β€’ Strong programming skills in Python for data manipulation and automation. β€’ Hands-on experience with PySpark for big data processing. β€’ Proficiency in AWS cloud services (S3, Glue, EMR, Redshift, Lambda, Athena, CloudWatch, etc.). β€’ Experience with ETL/ELT workflows and data pipeline orchestration tools (e.g., Airflow, Step Functions). β€’ Solid understanding of data modeling, warehousing, and data lake concepts. β€’ Knowledge of SQL and experience working with relational and NoSQL databases. β€’ Familiarity with CI/CD, version control (Git), and Agile development practices. β€’ Strong problem-solving skills and the ability to work independently or within a team. Preferred Qualifications (Nice To Have) β€’ Experience with containerization tools (Docker, Kubernetes). β€’ Exposure to streaming technologies (Kafka, Kinesis). β€’ Knowledge of data governance, security, and compliance in cloud environments. β€’ Familiarity with BI/Analytics tools (Tableau, Power BI, QuickSight). Education Bachelors or Masters degree in Computer Science, Information Technology, Data Engineering, or a related field.