

AI Cloud Engineer (LLama/OPenAI/RAG) _ Only W2
β - Featured Role | Apply direct with Data Freelance Hub
This role is for an AI Cloud Engineer in Charlotte, NC, for 12+ months at a W2 pay rate. Key skills include hybrid cloud solutions, RESTful APIs, generative AI optimization, Terraform, and Apache workflows. Experience with ML pipelines and observability tools is required.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
July 1, 2025
π - Project duration
More than 6 months
-
ποΈ - Location type
On-site
-
π - Contract type
W2 Contractor
-
π - Security clearance
Unknown
-
π - Location detailed
Charlotte, NC
-
π§ - Skills detailed
#Airflow #AWS (Amazon Web Services) #GCP (Google Cloud Platform) #ML (Machine Learning) #Batch #FastAPI #Datadog #Swagger #Cloud #Observability #Ansible #Spark (Apache Spark) #Apache Airflow #Splunk #Monitoring #AI (Artificial Intelligence) #Terraform #MLflow #Databases #Deployment #Scala #Azure #Debugging #Logging #Prometheus #Apache Iceberg
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
AI Cloud Engineer (LLama/OPenAI/RAG)
Location: charlotte, NC
Duration: 12 Months +
Only w2
We are looking for an experienced engineer with the following qualifications:
β’ Expertise in designing and implementing hybrid cloud solutions across AWS, GCP, and Azure, ensuring availability, scalability, and cost efficiency.
β’ Proficiency in building RESTful APIs using FastAPI and Swagger for real-time LLM inference, including scalable model serving pipelines.
β’ Experience optimizing generative AI models (LLaMA, Mistral, OpenAI GPT) and implementing RAG pipelines with Ray and VectorAI for distributed, context-aware inferencing.
β’ Strong skills in automating infrastructure with Terraform, Ansible, and Crossplane for multi-cloud deployments.
β’ Familiarity with MLflow, DVC, and VectorAI to support reproducible and scalable ML pipelines.
β’ Ability to provision GPU-accelerated infrastructure to boost LLM training performance by up to 50%.
β’ Experience using Apache Iceberg with vector databases (Milvus, Pinecone) for semantic search and dataset lineage.
β’ Skilled in orchestrating real-time and batch data workflows using Apache Airflow, Spark, and Flink.
β’ Knowledge of observability tools like Prometheus, Datadog, and Splunk for monitoring, logging, and alerting.
β’ Capable of designing dashboards and metrics pipelines to deliver insights and reduce debugging time.