

Full-Stack GenAI Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Full-Stack GenAI Engineer, onsite, with a contract length of "unknown" and a pay rate of "unknown." Key skills include Python, LangChain, AWS, and Azure. A Master's degree and 3+ years in AI development are preferred.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
September 18, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Iselin, NJ
-
π§ - Skills detailed
#FastAPI #Automated Testing #Monitoring #React #Databases #AI (Artificial Intelligence) #PostgreSQL #Data Processing #Langchain #Azure DevOps #Scala #"ETL (Extract #Transform #Load)" #Spark (Apache Spark) #DevOps #PySpark #TypeScript #Streamlit #AWS (Amazon Web Services) #ADF (Azure Data Factory) #Redshift #Terraform #.Net #OpenCV (Open Source Computer Vision Library) #Python #Azure #SageMaker #ML (Machine Learning) #C++ #Cloud #Programming #MongoDB #Django #Microsoft Azure #AWS SageMaker #Azure Data Factory #PyTorch #Airflow #AWS Glue #Computer Science
Role description
Job Title: Full-Stack GenAI Engineer
Location: Onsite
Overview:
We are seeking a highly skilled and innovative Full-Stack GenAI Engineer to join our clientβs AI team. The ideal candidate will have hands-on experience building end-to-end AI solutions, including LLM-based applications, retrieval-augmented generation (RAG) systems, and scalable infrastructure. This role requires a blend of software engineering, machine learning, and prompt engineering expertise to deliver impactful AI products.
Key Responsibilities:
β’ Design and deploy agentic LLM systems using LangChain, FastAPI, and front-end frameworks.
β’ Optimize model performance through quantization (e.g., 4-bit LLaMA models) and latency improvements.
β’ Build and maintain RAG pipelines using hybrid sparse-dense retrieval, semantic compression, and chunk-chaining.
β’ Fine-tune prompts and client queries using DSPy, adapters, and prompt engineering techniques.
β’ Develop and deploy ML models using tools like Kubeflow, Airflow, and Azure/AWS infrastructure.
β’ Create and manage ETL pipelines for structured and unstructured data using MongoDB, PostgreSQL, AWS Glue, and Azure Data Factory.
β’ Mentor junior engineers and contribute to open-source projects and internal research initiatives.
β’ Integrate telemetry and automated testing for robust AI system monitoring and validation.
Required Skills & Technologies:
β’ Programming & Frameworks: Python, PySpark, .NET, Django, FastAPI, NextJS, TypeScript, React
β’ AI/ML Tools: LangChain, PyTorch, Huggingface, OpenCV, DSPy, Whisper.cpp, Granite adapters
β’ Cloud & DevOps: AWS (Sagemaker, Glue, Redshift), Azure (OpenAI, Bot Framework, LUIS), Terraform, CloudFormation, Azure DevOps
β’ Databases: PostgreSQL, MongoDB, Pinecone
β’ Certifications: Microsoft Azure Fundamentals (AZ-900) preferred
β’ Other Tools: Kubeflow, Airflow, Ngrok, Streamlit, OCR (pytesseract), VLM APIs
Preferred Qualifications:
β’ Masterβs degree in Computer Science or related field
β’ 3+ years of professional experience in software engineering or AI development
β’ Experience with multimodal data processing and retrieval systems
β’ Strong communication skills and ability to engage with clients and stakeholders
β’ Demonstrated success in hackathons, open-source contributions, or research projects
Job Title: Full-Stack GenAI Engineer
Location: Onsite
Overview:
We are seeking a highly skilled and innovative Full-Stack GenAI Engineer to join our clientβs AI team. The ideal candidate will have hands-on experience building end-to-end AI solutions, including LLM-based applications, retrieval-augmented generation (RAG) systems, and scalable infrastructure. This role requires a blend of software engineering, machine learning, and prompt engineering expertise to deliver impactful AI products.
Key Responsibilities:
β’ Design and deploy agentic LLM systems using LangChain, FastAPI, and front-end frameworks.
β’ Optimize model performance through quantization (e.g., 4-bit LLaMA models) and latency improvements.
β’ Build and maintain RAG pipelines using hybrid sparse-dense retrieval, semantic compression, and chunk-chaining.
β’ Fine-tune prompts and client queries using DSPy, adapters, and prompt engineering techniques.
β’ Develop and deploy ML models using tools like Kubeflow, Airflow, and Azure/AWS infrastructure.
β’ Create and manage ETL pipelines for structured and unstructured data using MongoDB, PostgreSQL, AWS Glue, and Azure Data Factory.
β’ Mentor junior engineers and contribute to open-source projects and internal research initiatives.
β’ Integrate telemetry and automated testing for robust AI system monitoring and validation.
Required Skills & Technologies:
β’ Programming & Frameworks: Python, PySpark, .NET, Django, FastAPI, NextJS, TypeScript, React
β’ AI/ML Tools: LangChain, PyTorch, Huggingface, OpenCV, DSPy, Whisper.cpp, Granite adapters
β’ Cloud & DevOps: AWS (Sagemaker, Glue, Redshift), Azure (OpenAI, Bot Framework, LUIS), Terraform, CloudFormation, Azure DevOps
β’ Databases: PostgreSQL, MongoDB, Pinecone
β’ Certifications: Microsoft Azure Fundamentals (AZ-900) preferred
β’ Other Tools: Kubeflow, Airflow, Ngrok, Streamlit, OCR (pytesseract), VLM APIs
Preferred Qualifications:
β’ Masterβs degree in Computer Science or related field
β’ 3+ years of professional experience in software engineering or AI development
β’ Experience with multimodal data processing and retrieval systems
β’ Strong communication skills and ability to engage with clients and stakeholders
β’ Demonstrated success in hackathons, open-source contributions, or research projects