Full-Stack GenAI Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Full-Stack GenAI Engineer, onsite, with a contract length of "unknown" and a pay rate of "unknown." Key skills include Python, LangChain, AWS, and Azure. A Master's degree and 3+ years in AI development are preferred.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
-
πŸ—“οΈ - Date discovered
September 18, 2025
πŸ•’ - Project duration
Unknown
-
🏝️ - Location type
On-site
-
πŸ“„ - Contract type
Unknown
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
Iselin, NJ
-
🧠 - Skills detailed
#FastAPI #Automated Testing #Monitoring #React #Databases #AI (Artificial Intelligence) #PostgreSQL #Data Processing #Langchain #Azure DevOps #Scala #"ETL (Extract #Transform #Load)" #Spark (Apache Spark) #DevOps #PySpark #TypeScript #Streamlit #AWS (Amazon Web Services) #ADF (Azure Data Factory) #Redshift #Terraform #.Net #OpenCV (Open Source Computer Vision Library) #Python #Azure #SageMaker #ML (Machine Learning) #C++ #Cloud #Programming #MongoDB #Django #Microsoft Azure #AWS SageMaker #Azure Data Factory #PyTorch #Airflow #AWS Glue #Computer Science
Role description
Job Title: Full-Stack GenAI Engineer Location: Onsite Overview: We are seeking a highly skilled and innovative Full-Stack GenAI Engineer to join our client’s AI team. The ideal candidate will have hands-on experience building end-to-end AI solutions, including LLM-based applications, retrieval-augmented generation (RAG) systems, and scalable infrastructure. This role requires a blend of software engineering, machine learning, and prompt engineering expertise to deliver impactful AI products. Key Responsibilities: β€’ Design and deploy agentic LLM systems using LangChain, FastAPI, and front-end frameworks. β€’ Optimize model performance through quantization (e.g., 4-bit LLaMA models) and latency improvements. β€’ Build and maintain RAG pipelines using hybrid sparse-dense retrieval, semantic compression, and chunk-chaining. β€’ Fine-tune prompts and client queries using DSPy, adapters, and prompt engineering techniques. β€’ Develop and deploy ML models using tools like Kubeflow, Airflow, and Azure/AWS infrastructure. β€’ Create and manage ETL pipelines for structured and unstructured data using MongoDB, PostgreSQL, AWS Glue, and Azure Data Factory. β€’ Mentor junior engineers and contribute to open-source projects and internal research initiatives. β€’ Integrate telemetry and automated testing for robust AI system monitoring and validation. Required Skills & Technologies: β€’ Programming & Frameworks: Python, PySpark, .NET, Django, FastAPI, NextJS, TypeScript, React β€’ AI/ML Tools: LangChain, PyTorch, Huggingface, OpenCV, DSPy, Whisper.cpp, Granite adapters β€’ Cloud & DevOps: AWS (Sagemaker, Glue, Redshift), Azure (OpenAI, Bot Framework, LUIS), Terraform, CloudFormation, Azure DevOps β€’ Databases: PostgreSQL, MongoDB, Pinecone β€’ Certifications: Microsoft Azure Fundamentals (AZ-900) preferred β€’ Other Tools: Kubeflow, Airflow, Ngrok, Streamlit, OCR (pytesseract), VLM APIs Preferred Qualifications: β€’ Master’s degree in Computer Science or related field β€’ 3+ years of professional experience in software engineering or AI development β€’ Experience with multimodal data processing and retrieval systems β€’ Strong communication skills and ability to engage with clients and stakeholders β€’ Demonstrated success in hackathons, open-source contributions, or research projects