GenAI Engineer (LLM Engineer)

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a GenAI Engineer (LLM Engineer) with a contract length of over 1 year, offering a remote position for candidates local to the Bay Area. Key skills include deep learning, NLP, and expertise in LLM optimization techniques.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
-
πŸ—“οΈ - Date discovered
August 15, 2025
πŸ•’ - Project duration
More than 6 months
-
🏝️ - Location type
Remote
-
πŸ“„ - Contract type
Unknown
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
San Francisco Bay Area
-
🧠 - Skills detailed
#Hugging Face #AI (Artificial Intelligence) #Mathematics #Documentation #Databases #Compliance #Data Science #Deep Learning #TensorFlow #"ETL (Extract #Transform #Load)" #PyTorch #NLP (Natural Language Processing) #Monitoring #Data Pipeline #Libraries #Transformers
Role description
We are looking for GenAI/LLM Engineer Remote(Local to bay area) 1 Year+ Implementing GenAI requires specialized expertise in large language models. Traditional data scientists often haven't had the opportunity to dive deep into the practical intricacies of LLMsβ€”particularly advanced fine-tuning techniques, model compression strategies, memory optimization approaches, and specialized training workflows. This role requires a hands-on deep learning practitioner comfortable with modern frameworks and libraries specific to LLM development. β€’ Enables domain-specific fine-tuning of models to unique utility context β€’ Improves model performance while reducing computational costs through advanced optimization techniques β€’ Creates specific AI capabilities that address our unique operational challenges β€’ Enables the CoE to move beyond generic AI tools to customized solutions that deliver higher business value Key Responsibilities: β€’ Implement and optimize advanced fine-tuning approaches (LoRA, PEFT, QLoRA) to adapt foundation models β€’ Develop systematic prompt engineering methodologies specific to utility operations, regulatory compliance, and technical documentation β€’ Create reusable prompt templates and libraries to standardize interactions across multiple LLM applications and use cases β€’ Implement prompt testing frameworks to quantitatively evaluate and iteratively improve prompt effectiveness β€’ Establish prompt versioning systems and governance to maintain consistency and quality across applications β€’ Apply model customization techniques like knowledge distillation, quantization, and pruning to reduce memory footprint and inference costs β€’ Tackle memory constraints using techniques such as sharded data parallelism, GPU offloading, or CPU+GPU hybrid approaches β€’ Build robust retrieval-augmented generation (RAG) pipelines with vector databases, embedding pipelines, and optimized chunking strategies β€’ Design advanced prompting strategies including chain-of-thought reasoning, conversation orchestration, and agent-based approaches β€’ Collaborate with the MLOps engineer to ensure models are efficiently deployed, monitored, and retrained as needed Expected Skillset: β€’ Deep Learning & NLP: Proficiency with PyTorch/TensorFlow, Hugging Face Transformers, DSPy, and advanced LLM training techniques β€’ GPU/Hardware Knowledge: Experience with multi-GPU training, memory optimization, and parallelization strategies β€’ LLMOps: Familiarity with workflows for maintaining LLM-based applications in production and monitoring model performance β€’ Technical Adaptability: Ability to interpret research papers and implement emerging techniques (without necessarily requiring PhD-level mathematics) β€’ Domain Adaptation: Skills in creating data pipelines for fine-tuning models with utility-specific content