

Agentic Data Engineer (Must Have AI, GIS, Azure Databricks, LLM, ETL and ELT Pipelines
β - Featured Role | Apply direct with Data Freelance Hub
This role is for an Agentic Data Engineer in Richmond, VA, with a contract length of "Quarterly" and a pay rate of "unknown." Key skills include AI, GIS, Azure Databricks, ETL/ELT pipelines, and experience with big data frameworks.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
July 10, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Richmond, VA
-
π§ - Skills detailed
#Data Engineering #Data Framework #Datasets #"ETL (Extract #Transform #Load)" #Databases #Azure Databricks #Data Quality #Data Architecture #Spark (Apache Spark) #Programming #Data Lake #Graph Databases #ML (Machine Learning) #Big Data #Cloud #AI (Artificial Intelligence) #Azure #Data Science #Python #GIT #Data Pipeline #Computer Science #Databricks #Storage #Apache Spark #Spatial Data #Data Storage #Version Control #Azure Machine Learning
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Resource will need to be in Richmond, VA Quarterly.
Both Phone and In Person Interview is required
The Virginia Department of Transportation's Information Technology Division is seeking a highly skilled Agentic Engineer to design, develop that solve real-world problems. The ideal candidate will have experience in designing data process to support agentic systems. ensure data quality and facilitating interaction between agents and data.
Responsibilities
β’ Design and develop data pipelines for agentic systems; develop robust data flows to handle complex interactions between AI agents and data sources.
β’ Train and fine-tune large language models (LLMs).
β’ Design and build data architecture, including databases and data lakes, to support various data engineering tasks.
β’ Develop and manage Extract, Load, Transform (ELT) processes to ensure data is accurately and efficiently moved from source systems to analytical platforms used in data science.
β’ Implement data pipelines that facilitate feedback loops, allowing human input to improve system performance in human-in-the-loop systems.
β’ Work with vector databases to store and retrieve embeddings efficiently.
β’ Collaborate with data scientists and engineers to preprocess data, train models, and integrate AI into applications.
β’ Optimize data storage and retrieval for high performance.
β’ Conduct statistical analysis to identify trends and patterns and create data formats from multiple sources.
Qualifications
β’ Strong data engineering fundamentals.
β’ Experience with big data frameworks such as Apache Spark and Azure Databricks.
β’ Ability to train LLMs using structured and unstructured datasets.
β’ Understanding of graph databases.
β’ Experience with Azure services including Blob Storage, Data Lakes, Databricks, Azure Machine Learning, Azure Computer Vision, Azure Video Indexer, Azure OpenAI, Azure Media Services, and Azure AI Search.
β’ Proficient in determining effective data partitioning criteria and implementing partition schemas using Spark.
β’ Understanding of core machine learning concepts and algorithms.
β’ Familiarity with cloud computing concepts and practices.
β’ Strong programming skills in Python and experience with AI/ML frameworks.
β’ Proficiency in working with vector databases and embedding models for retrieval tasks.
β’ Expertise in integrating with AI agent frameworks.
β’ Experience with cloud-based AI services, especially Azure AI.
β’ Proficient with version control systems such as Git.
β’ Bachelorβs or Masterβs degree in Computer Science, Artificial Intelligence, Data Science, or a related field
Required Skills
β’ Understanding of Big Data Technologies β 5 Years
β’ Experience developing ETL and ELT pipelines β 5 Years
β’ Experience with Spark, GraphDB, Azure Databricks β 5 Years
β’ Experience training LLMs with structured and unstructured data sets β 4 Years
β’ Experience in Data Partitioning and Data Conflation β 3 Years
β’ Experience with GIS spatial data β 3 Years