Anblicks

Technology and Operations - Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with 7+ years of experience, focusing on big data technologies and AI/ML systems. Located in Sterling, it offers a competitive pay rate for a contract length of "X months." Key skills include Scala, Python, Apache Spark, and cloud platforms.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
December 23, 2025
πŸ•’ - Duration
Unknown
-
🏝️ - Location
On-site
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Virginia, United States
-
🧠 - Skills detailed
#"ETL (Extract #Transform #Load)" #Schema Design #Cloud #Spark (Apache Spark) #Apache Spark #Data Mining #Python #Data Accuracy #Redshift #SQL (Structured Query Language) #Kafka (Apache Kafka) #Apache Airflow #Data Processing #Big Data #Tableau #Libraries #Batch #AI (Artificial Intelligence) #Data Analysis #Documentation #PyTorch #Programming #Scala #Data Access #GitHub #Databricks #Looker #Java #Snowflake #Data Engineering #Data Pipeline #ML (Machine Learning) #Data Quality #TensorFlow #Airflow #Databases #Leadership #Computer Science #AWS (Amazon Web Services)
Role description
Job Title: Technology and Operations - Data Engineer Location: Sterling - 45580 Terminal Drive\_US28-Virginia We are looking for a seasoned Data Engineer with a strong foundation in big data technologies and a growing proficiency in AI/ML systems. This individual will bring deep expertise in large-scale data processing frameworks (both open-source and proprietary), OLAP/OLTP systems, and real-time data streaming. The ideal candidate will also demonstrate a passion for enabling AI-driven solutions through robust, scalable data infrastructure. Key Responsibilities: β€’ Design, develop, and maintain highly scalable, fault-tolerant real-time, near real-time, and batch data pipelines. β€’ Implement data quality checks, validation, and cleaning processes to ensure high data accuracy and integrity. β€’ Continuously monitor and optimize data pipelines and databases for performance, resource utilization, and cost efficiency. β€’ Uphold high standards in code quality, testing, and documentation. β€’ Mentor junior data engineers and provide technical leadership within the team. β€’ Perform exploratory and quantitative analytics, data mining, and discovery to support AI/ML initiatives. β€’ Collaborate with data analysts and business stakeholders to make data accessible and actionable. β€’ Participate in 24x7 platform support rotations as needed. Required Qualifications: β€’ Bachelor’s degree or higher in Computer Science or a related field. β€’ 7+ years of experience in data engineering roles. β€’ Proficiency in programming languages such as Scala, Python, or Java. β€’ Expertise in distributed data processing frameworks like Apache Spark or Flink. β€’ Experience with stream processing systems such as Kafka or Kinesis. β€’ Strong knowledge of cloud platforms (e.g., AWS) and cloud-native data platforms like Databricks, Snowflake, or Redshift. β€’ Solid understanding of SQL, schema design, dimensional modeling, and ETL best practices. β€’ Experience with workflow orchestration tools such as Apache Airflow. β€’ Familiarity with CI/CD pipelines, preferably GitHub Actions. β€’ Strong technical documentation and issue-tracking skills. β€’ Experience with analytics tools such as Looker or Tableau. β€’ Demonstrated ability to adopt and integrate new technologies. β€’ Experience in direct-to-consumer digital businesses is a plus. Preferred Qualifications: β€’ Familiarity with statistical modeling libraries (e.g., Scikit-learn, XGBoost, TensorFlow). β€’ Experience with AI/ML frameworks such as TensorFlow, PyTorch, or Scikit-learn. β€’ Exposure to GenAI technologies, including LLMs and RAG pipelines.