W3Global

Senior Full Stack GenAI Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Full Stack GenAI Engineer, offering a 6-12 month hybrid contract in Plano, TX. Pay rate is unspecified. Requires 10+ years in software engineering, 5+ years in full-stack (Python/Java), and 3+ years with GenAI tools.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
November 5, 2025
πŸ•’ - Duration
More than 6 months
-
🏝️ - Location
Hybrid
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Plano, TX
-
🧠 - Skills detailed
#Databases #AWS (Amazon Web Services) #Docker #ML (Machine Learning) #Automation #AI (Artificial Intelligence) #MLflow #Scala #Langchain #Kubernetes #Python #Deployment #Cloud #Java #Microservices #SageMaker #Strategy #Hugging Face
Role description
Our client is seeking a Senior Full Stack GenAI Engineer to help design, build, and deploy next-generation AI-driven applications. This hands-on role combines software engineering, architecture, and applied AI expertise to create scalable, secure, and modern solutions leveraging GenAI tools, LLMs, and AWS services. The ideal candidate will have a strong full-stack foundation in Python and Java, extensive experience with microservices and CI/CD pipelines, and a passion for developing innovative AI-powered systems. This position offers the opportunity to shape the client's AI strategy and influence platform direction through experimentation, prototyping, and deployment of real-world GenAI applications. Key Responsibilities β€’ Design and implement scalable GenAI architectures using large language models (LLMs), multimodal models, and retrieval-augmented generation (RAG). β€’ Develop and deploy AI-powered applications leveraging Python/Java, microservices, and cloud-native design patterns. β€’ Integrate AI and GenAI capabilities into enterprise applications using APIs, SDKs, and orchestration frameworks such as LangChain and Bedrock. β€’ Build and maintain CI/CD pipelines for AI and full-stack applications, ensuring automation, versioning, and efficient deployment. β€’ Fine-tune and optimize models using techniques such as instruction tuning, RLHF, quantization, and distillation. β€’ Experiment with emerging GenAI tools and frameworks (e.g., Hugging Face, LlamaIndex, DeepEval, RAGAS). β€’ Implement responsible AI practices, including bias detection, hallucination mitigation, and explainability. β€’ Collaborate with cross-functional product and architecture teams to align technical decisions with business goals. β€’ Stay ahead of industry trends and drive innovation within the GenAI and full-stack development ecosystem. Required Qualifications β€’ 10+ years of total software engineering experience, including 5+ years full-stack development (Python/Java). β€’ 3+ years working directly with GenAI frameworks and tools (e.g., LangChain, Hugging Face, Bedrock, vector databases). β€’ Hands-on experience building and deploying GenAI applications in production environments. β€’ Strong background in modern architecture styles - microservices, APIs, and event-driven systems. β€’ Expertise with AWS services (SageMaker, Bedrock, Comprehend, Lex) or equivalent cloud platforms. β€’ Experience with CI/CD pipelines, containerization (Docker, Kubernetes), and ML workflow tools (MLflow, DVC). β€’ Strong understanding of LLMs, embeddings, and vector databases (e.g., FAISS, Pinecone, Weaviate). β€’ Excellent communication and problem-solving skills with the ability to work collaboratively in cross-functional teams. Preferred Qualifications β€’ AWS AI/ML certification. β€’ Familiarity with AI testing frameworks such as DeepEval or RAGAS. β€’ Financial services or large enterprise environment experience. β€’ Knowledge of prompt engineering, few-shot learning, and agent-based systems. Role Details β€’ Location: Hybrid (3 days onsite in Plano, TX - Tuesday through Thursday) β€’ Contract: 6-12 months (possibility of extension or conversion) β€’ Interview Process: β€’ MS Teams - Technical Screening (30 min) β€’ Onsite - 1 hour (includes coding)