

Anblicks
Technology and Operations - Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with 7+ years of experience, focusing on big data technologies and AI/ML systems. Located in Sterling, it offers a competitive pay rate for a contract length of "X months." Key skills include Scala, Python, Apache Spark, and cloud platforms.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
December 23, 2025
π - Duration
Unknown
-
ποΈ - Location
On-site
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Virginia, United States
-
π§ - Skills detailed
#"ETL (Extract #Transform #Load)" #Schema Design #Cloud #Spark (Apache Spark) #Apache Spark #Data Mining #Python #Data Accuracy #Redshift #SQL (Structured Query Language) #Kafka (Apache Kafka) #Apache Airflow #Data Processing #Big Data #Tableau #Libraries #Batch #AI (Artificial Intelligence) #Data Analysis #Documentation #PyTorch #Programming #Scala #Data Access #GitHub #Databricks #Looker #Java #Snowflake #Data Engineering #Data Pipeline #ML (Machine Learning) #Data Quality #TensorFlow #Airflow #Databases #Leadership #Computer Science #AWS (Amazon Web Services)
Role description
Job Title: Technology and Operations - Data Engineer
Location: Sterling - 45580 Terminal Drive\_US28-Virginia
We are looking for a seasoned Data Engineer with a strong foundation in big data technologies and a growing proficiency in AI/ML systems. This individual will bring deep expertise in large-scale data processing frameworks (both open-source and proprietary), OLAP/OLTP systems, and real-time data streaming. The ideal candidate will also demonstrate a passion for enabling AI-driven solutions through robust, scalable data infrastructure.
Key Responsibilities:
β’ Design, develop, and maintain highly scalable, fault-tolerant real-time, near real-time, and batch data pipelines.
β’ Implement data quality checks, validation, and cleaning processes to ensure high data accuracy and integrity.
β’ Continuously monitor and optimize data pipelines and databases for performance, resource utilization, and cost efficiency.
β’ Uphold high standards in code quality, testing, and documentation.
β’ Mentor junior data engineers and provide technical leadership within the team.
β’ Perform exploratory and quantitative analytics, data mining, and discovery to support AI/ML initiatives.
β’ Collaborate with data analysts and business stakeholders to make data accessible and actionable.
β’ Participate in 24x7 platform support rotations as needed.
Required Qualifications:
β’ Bachelorβs degree or higher in Computer Science or a related field.
β’ 7+ years of experience in data engineering roles.
β’ Proficiency in programming languages such as Scala, Python, or Java.
β’ Expertise in distributed data processing frameworks like Apache Spark or Flink.
β’ Experience with stream processing systems such as Kafka or Kinesis.
β’ Strong knowledge of cloud platforms (e.g., AWS) and cloud-native data platforms like Databricks, Snowflake, or Redshift.
β’ Solid understanding of SQL, schema design, dimensional modeling, and ETL best practices.
β’ Experience with workflow orchestration tools such as Apache Airflow.
β’ Familiarity with CI/CD pipelines, preferably GitHub Actions.
β’ Strong technical documentation and issue-tracking skills.
β’ Experience with analytics tools such as Looker or Tableau.
β’ Demonstrated ability to adopt and integrate new technologies.
β’ Experience in direct-to-consumer digital businesses is a plus.
Preferred Qualifications:
β’ Familiarity with statistical modeling libraries (e.g., Scikit-learn, XGBoost, TensorFlow).
β’ Experience with AI/ML frameworks such as TensorFlow, PyTorch, or Scikit-learn.
β’ Exposure to GenAI technologies, including LLMs and RAG pipelines.
Job Title: Technology and Operations - Data Engineer
Location: Sterling - 45580 Terminal Drive\_US28-Virginia
We are looking for a seasoned Data Engineer with a strong foundation in big data technologies and a growing proficiency in AI/ML systems. This individual will bring deep expertise in large-scale data processing frameworks (both open-source and proprietary), OLAP/OLTP systems, and real-time data streaming. The ideal candidate will also demonstrate a passion for enabling AI-driven solutions through robust, scalable data infrastructure.
Key Responsibilities:
β’ Design, develop, and maintain highly scalable, fault-tolerant real-time, near real-time, and batch data pipelines.
β’ Implement data quality checks, validation, and cleaning processes to ensure high data accuracy and integrity.
β’ Continuously monitor and optimize data pipelines and databases for performance, resource utilization, and cost efficiency.
β’ Uphold high standards in code quality, testing, and documentation.
β’ Mentor junior data engineers and provide technical leadership within the team.
β’ Perform exploratory and quantitative analytics, data mining, and discovery to support AI/ML initiatives.
β’ Collaborate with data analysts and business stakeholders to make data accessible and actionable.
β’ Participate in 24x7 platform support rotations as needed.
Required Qualifications:
β’ Bachelorβs degree or higher in Computer Science or a related field.
β’ 7+ years of experience in data engineering roles.
β’ Proficiency in programming languages such as Scala, Python, or Java.
β’ Expertise in distributed data processing frameworks like Apache Spark or Flink.
β’ Experience with stream processing systems such as Kafka or Kinesis.
β’ Strong knowledge of cloud platforms (e.g., AWS) and cloud-native data platforms like Databricks, Snowflake, or Redshift.
β’ Solid understanding of SQL, schema design, dimensional modeling, and ETL best practices.
β’ Experience with workflow orchestration tools such as Apache Airflow.
β’ Familiarity with CI/CD pipelines, preferably GitHub Actions.
β’ Strong technical documentation and issue-tracking skills.
β’ Experience with analytics tools such as Looker or Tableau.
β’ Demonstrated ability to adopt and integrate new technologies.
β’ Experience in direct-to-consumer digital businesses is a plus.
Preferred Qualifications:
β’ Familiarity with statistical modeling libraries (e.g., Scikit-learn, XGBoost, TensorFlow).
β’ Experience with AI/ML frameworks such as TensorFlow, PyTorch, or Scikit-learn.
β’ Exposure to GenAI technologies, including LLMs and RAG pipelines.





