Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with a 3-month contract, paying $50-$55 per hour, remote in Lake Bluff, IL. Key skills include Azure PostgreSQL, ETL/ELT pipeline development, Python proficiency, and OCR tool familiarity. Experience with large-scale contract ingestion is preferred.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
440
-
πŸ—“οΈ - Date discovered
September 27, 2025
πŸ•’ - Project duration
3 to 6 months
-
🏝️ - Location type
Remote
-
πŸ“„ - Contract type
W2 Contractor
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
United States
-
🧠 - Skills detailed
#AI (Artificial Intelligence) #"ETL (Extract #Transform #Load)" #Agile #Scala #Data Pipeline #Scripting #SAP #Data Extraction #Azure #Databases #PostgreSQL #Python #Security #Automation #Data Engineering #Documentation #Database Schema #ML (Machine Learning) #API (Application Programming Interface) #Schema Design #GitHub #REST (Representational State Transfer) #Metadata
Role description
Job Title: Data Engineer Contract Duration: 3 Months Location: Lake Bluff, IL Work Arrangement: Remote Pay Range: $50 - $55 per hour β€’ β€’ β€’ Note: No C2C or C2H applicable ROLE OVERVIEW As our Data Engineer, you will own the end-to-end data pipelines. This includes designing scalable databases, developing ingestion workflows, collaborating with our internal Machine Learning Engineering team, and structuring supplier spend data. You’ll work closely with the Full Stack Developer to co-design the database schema for the Negotiation AI and ensure future compatibility with the ingestion pipeline. KEY DELIVERABLES β€’ Ingestion Pipeline: Build and deploy a robust ETL/ELT pipeline using Azure to ingest 50,000+ contracts. β€’ Metadata Extraction: Configure and run OCR workflows (e.g., OlmOCR/Azure Document Intelligence) to extract key contract fields such as dates, parties, terms etc. β€’ Scalable Database Schema: Design and implement a schema in Azure PostgreSQL to store contract metadata, OCR outputs, and supplier spend data. Collaborate with the Software Developer to design a future-ready schema for AI consumption. REQUIRED SKILLS & EXPERIENCE Data Engineering & ETL/ELT β€’ Experience with Azure PostgreSQL or similar relational databases β€’ Skilled in building scalable ETL/ELT pipelines (preferably using Azure) β€’ Proficient in Python for scripting and automation OCR Collaboration β€’ Ability to work with internal Machine Learning Engineering teams to validate and structure extracted data β€’ Familiarity with OCR tools (e.g., Azure Document Intelligence, Tesseract) is a plus SAP Ariba Integration β€’ Exposure to cXML, ARBCI, SOAP/REST protocols is a plus β€’ Comfortable with API authentication (OAuth, tokens) and enterprise-grade security Agile Collaboration & Documentation β€’ Comfortable working in sprints and cross-functional teams β€’ Able to use Github Copilot to document practices for handover PREFERRED QUALIFICATIONS β€’ Experience with large-scale contract ingestion projects β€’ Familiarity with procurement systems and contract lifecycle management β€’ Background in integrating data pipelines with AI or analytics platforms WHY JOIN US? β€’ Focused Scope with Future Impact: Lay the foundation for an AI-driven negotiation platform β€’ Cutting-Edge Tools: Work with SAP Ariba, OCR, Azure, and advanced analytics β€’ Collaborative Environment: Partner with Software Developers and AI specialists