

Gen AI Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Gen AI Engineer with 8-15 years of AI development experience, offering a contract for 40 hours per week at $45.00 - $65.00 per hour. Located in Dallas, TX, key skills include deep learning, NLP, and proficiency in PyTorch/TensorFlow.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
520
-
🗓️ - Date discovered
August 29, 2025
🕒 - Project duration
Unknown
-
🏝️ - Location type
On-site
-
📄 - Contract type
Unknown
-
🔒 - Security clearance
Unknown
-
📍 - Location detailed
Dallas, TX 75211
-
🧠 - Skills detailed
#Compliance #AI (Artificial Intelligence) #Mathematics #"ETL (Extract #Transform #Load)" #Databases #Hugging Face #Transformers #Data Pipeline #Deep Learning #Monitoring #PyTorch #TensorFlow #Data Science #Documentation #NLP (Natural Language Processing) #Libraries
Role description
Job Summary
Job Description:
Looking for someone who can not only code but also be a lead and liaison to the client's leaders and work with a TPM and an Analyst.
Please find someone with 8-15 years of AI development experience.
KEY POINTS:
MUST be able to drive initiatives
MUST be able to determine what the business is looking for and build a POC
MUST have “Executive Presence” and be delightful too.
RE: Data Science: MUST have the ability to adapt and understand ‘use cases’
Locals in PST only.
GenAI/LLM Engineer (NLP, TensorFlow, PyTorch SME)
Implementing GenAI requires specialized expertise in large language models. Traditional data scientists often haven't had the opportunity to dive deep into the practical intricacies of LLMs—particularly advanced fine-tuning techniques, model compression strategies, memory optimization approaches, and specialized training workflows. This role requires a hands-on deep learning practitioner comfortable with modern frameworks and libraries specific to LLM development.
Enables domain-specific fine-tuning of models to client's unique utility context
Improves model performance while reducing computational costs through advanced optimization techniques
Creates Client-specific AI capabilities that address our unique operational challenges
Enables the CoE to move beyond generic AI tools to customized solutions that deliver higher business value
Key Responsibilities:
Implement and optimize advanced fine-tuning approaches (LoRA, PEFT, QLoRA) to adapt foundation models to client's domain
Develop systematic prompt engineering methodologies specific to utility operations, regulatory compliance, and technical documentation
Create reusable prompt templates and libraries to standardize interactions across multiple LLM applications and use cases
Implement prompt testing frameworks to quantitatively evaluate and iteratively improve prompt effectiveness
Establish prompt versioning systems and governance to maintain consistency and quality across applications
Apply model customization techniques like knowledge distillation, quantization, and pruning to reduce memory footprint and inference costs
Tackle memory constraints using techniques such as sharded data parallelism, GPU offloading, or CPU+GPU hybrid approaches
Build robust retrieval-augmented generation (RAG) pipelines with vector databases, embedding pipelines, and optimized chunking strategies
Design advanced prompting strategies including chain-of-thought reasoning, conversation orchestration, and agent-based approaches
Collaborate with the MLOps engineer to ensure models are efficiently deployed, monitored, and retrained as needed
Expected Skillset:
Deep Learning & NLP: Proficiency with PyTorch/TensorFlow, Hugging Face Transformers, DSPy, and advanced LLM training techniques
GPU/Hardware Knowledge: Experience with multi-GPU training, memory optimization, and parallelization strategies
LLMOps: Familiarity with workflows for maintaining LLM-based applications in production and monitoring model performance
Technical Adaptability: Ability to interpret research papers and implement emerging techniques (without necessarily requiring PhD-level mathematics)
Domain Adaptation: Skills in creating data pipelines for fine-tuning models with utility-specific content
Please share Your Resume: James@allobr.com
Job Type: Contract
Pay: $45.00 - $65.00 per hour
Expected hours: 40 per week
Benefits:
Health insurance
Ability to Commute:
Dallas, TX 75211 (Required)
Ability to Relocate:
Dallas, TX 75211: Relocate before starting work (Required)
Work Location: In person
Job Summary
Job Description:
Looking for someone who can not only code but also be a lead and liaison to the client's leaders and work with a TPM and an Analyst.
Please find someone with 8-15 years of AI development experience.
KEY POINTS:
MUST be able to drive initiatives
MUST be able to determine what the business is looking for and build a POC
MUST have “Executive Presence” and be delightful too.
RE: Data Science: MUST have the ability to adapt and understand ‘use cases’
Locals in PST only.
GenAI/LLM Engineer (NLP, TensorFlow, PyTorch SME)
Implementing GenAI requires specialized expertise in large language models. Traditional data scientists often haven't had the opportunity to dive deep into the practical intricacies of LLMs—particularly advanced fine-tuning techniques, model compression strategies, memory optimization approaches, and specialized training workflows. This role requires a hands-on deep learning practitioner comfortable with modern frameworks and libraries specific to LLM development.
Enables domain-specific fine-tuning of models to client's unique utility context
Improves model performance while reducing computational costs through advanced optimization techniques
Creates Client-specific AI capabilities that address our unique operational challenges
Enables the CoE to move beyond generic AI tools to customized solutions that deliver higher business value
Key Responsibilities:
Implement and optimize advanced fine-tuning approaches (LoRA, PEFT, QLoRA) to adapt foundation models to client's domain
Develop systematic prompt engineering methodologies specific to utility operations, regulatory compliance, and technical documentation
Create reusable prompt templates and libraries to standardize interactions across multiple LLM applications and use cases
Implement prompt testing frameworks to quantitatively evaluate and iteratively improve prompt effectiveness
Establish prompt versioning systems and governance to maintain consistency and quality across applications
Apply model customization techniques like knowledge distillation, quantization, and pruning to reduce memory footprint and inference costs
Tackle memory constraints using techniques such as sharded data parallelism, GPU offloading, or CPU+GPU hybrid approaches
Build robust retrieval-augmented generation (RAG) pipelines with vector databases, embedding pipelines, and optimized chunking strategies
Design advanced prompting strategies including chain-of-thought reasoning, conversation orchestration, and agent-based approaches
Collaborate with the MLOps engineer to ensure models are efficiently deployed, monitored, and retrained as needed
Expected Skillset:
Deep Learning & NLP: Proficiency with PyTorch/TensorFlow, Hugging Face Transformers, DSPy, and advanced LLM training techniques
GPU/Hardware Knowledge: Experience with multi-GPU training, memory optimization, and parallelization strategies
LLMOps: Familiarity with workflows for maintaining LLM-based applications in production and monitoring model performance
Technical Adaptability: Ability to interpret research papers and implement emerging techniques (without necessarily requiring PhD-level mathematics)
Domain Adaptation: Skills in creating data pipelines for fine-tuning models with utility-specific content
Please share Your Resume: James@allobr.com
Job Type: Contract
Pay: $45.00 - $65.00 per hour
Expected hours: 40 per week
Benefits:
Health insurance
Ability to Commute:
Dallas, TX 75211 (Required)
Ability to Relocate:
Dallas, TX 75211: Relocate before starting work (Required)
Work Location: In person