

IMS Group
AI Specialist(MLops, Sage Maker - Only W2)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for an AI Specialist (MLOps, SageMaker) on a W2 contract for an unspecified duration, offering competitive pay. Key skills required include Python, MLOps, and AWS SageMaker, with a strong background in machine learning engineering and data science methodologies.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
February 27, 2026
π - Duration
Unknown
-
ποΈ - Location
Unknown
-
π - Contract
W2 Contractor
-
π - Security
Unknown
-
π - Location detailed
Charlotte, NC
-
π§ - Skills detailed
#Data Science #S3 (Amazon Simple Storage Service) #Observability #AWS S3 (Amazon Simple Storage Service) #Classification #GitHub #Regression #Python #Code Reviews #AWS (Amazon Web Services) #"ETL (Extract #Transform #Load)" #Monitoring #AI (Artificial Intelligence) #Libraries #Data Quality #AWS SageMaker #SSIS (SQL Server Integration Services) #Lambda (AWS Lambda) #Computer Science #Documentation #Pandas #Prometheus #EC2 #Big Data #SageMaker #Deployment #Version Control #Kubernetes #Sqoop (Apache Sqoop) #ML (Machine Learning) #NumPy #Agile #Programming #SQL (Structured Query Language) #Docker #Infrastructure as Code (IaC) #GIT #Spark (Apache Spark) #Terraform #Cloud #Hadoop
Role description
Summary:
We are seeking a skilled Machine Learning Engineer/AI Specialist to join our dynamic team. The ideal candidate will have extensive AWS SageMaker, strong Python programming skills, a solid background in data science, and a deep understanding of MLOps practices.
Top Skills: Python, MLOps, AWS SageMaker
Only W2, Not on C2C or 1099
Key Responsibilities:
β’ Design, develop, and deploy machine learning models using AWS SageMaker platform.
β’ Build and maintain ML pipelines for training, validation, and deployment of models.
β’ Implement MLOps best practices including CI/CD for machine learning workflows.
β’ Collaborate with data scientists to productionize research models.
β’ Monitor model performance and implement automated retraining processes.
β’ Optimize model inference performance and cost efficiency.
β’ Develop and maintain model versioning and experiment tracking systems.
β’ Ensure data quality and implement data validation frameworks.
β’ Create comprehensive documentation and technical specifications.
β’ Participate in code reviews and maintain high coding standards.
β’ Debug Terraform and Concourse errors.
β’ Proactively update pipelines based on changes made by other organizations.
β’ Migrate repository to GitHub and update pipelines accordingly.
Required Qualifications:
β’ Bachelor's degree in Computer Science, Data Science, Engineering, or related field; or 8 years of equivalent work experience.
β’ 3+ years of experience in machine learning engineering or related roles.
β’ Proficiency in Python programming with experience in ML libraries (pandas, numpy, etc.).
β’ Familiarity with Infrastructure as Code (Terraform, CloudFormation).
β’ Hands-on experience with AWS SageMaker for model training, tuning, and deployment.
β’ Strong background in data science methodologies and statistical analysis.
β’ Deep understanding of MLOps practices and tools (Docker, Kubernetes, CI/CD pipelines).
β’ Experience with version control systems (Git Hub Actions) and collaborative development.
β’ Knowledge of cloud platforms, preferably AWS (S3, EC2, Lambda, etc.).
Preferred Qualifications:
β’ Master's degree in a relevant field.
β’ AWS certifications (Machine Learning Specialty, Solutions Architect, etc.).
β’ Knowledge of containerization and orchestration technologies.
β’ Experience with monitoring and observability tools (CloudWatch, Prometheus, etc.).
β’ Experience with big data technologies (EMR, Spark, Hadoop, etc.).
β’ Understanding of software engineering best practices and design patterns.
β’ Good working experience in ETL (SSIS or Sqoop/Spark).
β’ Experience with EMR
β’ Expert SQL knowledge (All types of Joins, CTEβs, Indexes, Stored Procedures, SQL performance).
β’ Knowledge in building basic machine learning models (Classification & Regression).
β’ Knowledge in Docker/MLOps and its orchestrations.
Key Skills & Competencies:
β’ Strong analytical and problem-solving abilities.
β’ Excellent communication and collaboration skills.
β’ Ability to work in fast-paced, agile environments.
β’ Detail-oriented with a focus on code quality and documentation.
β’ Continuous learning mindset and adaptability to new technologies.
β’ Experience working cross-functionally with data scientists, engineers, and product teams.
Summary:
We are seeking a skilled Machine Learning Engineer/AI Specialist to join our dynamic team. The ideal candidate will have extensive AWS SageMaker, strong Python programming skills, a solid background in data science, and a deep understanding of MLOps practices.
Top Skills: Python, MLOps, AWS SageMaker
Only W2, Not on C2C or 1099
Key Responsibilities:
β’ Design, develop, and deploy machine learning models using AWS SageMaker platform.
β’ Build and maintain ML pipelines for training, validation, and deployment of models.
β’ Implement MLOps best practices including CI/CD for machine learning workflows.
β’ Collaborate with data scientists to productionize research models.
β’ Monitor model performance and implement automated retraining processes.
β’ Optimize model inference performance and cost efficiency.
β’ Develop and maintain model versioning and experiment tracking systems.
β’ Ensure data quality and implement data validation frameworks.
β’ Create comprehensive documentation and technical specifications.
β’ Participate in code reviews and maintain high coding standards.
β’ Debug Terraform and Concourse errors.
β’ Proactively update pipelines based on changes made by other organizations.
β’ Migrate repository to GitHub and update pipelines accordingly.
Required Qualifications:
β’ Bachelor's degree in Computer Science, Data Science, Engineering, or related field; or 8 years of equivalent work experience.
β’ 3+ years of experience in machine learning engineering or related roles.
β’ Proficiency in Python programming with experience in ML libraries (pandas, numpy, etc.).
β’ Familiarity with Infrastructure as Code (Terraform, CloudFormation).
β’ Hands-on experience with AWS SageMaker for model training, tuning, and deployment.
β’ Strong background in data science methodologies and statistical analysis.
β’ Deep understanding of MLOps practices and tools (Docker, Kubernetes, CI/CD pipelines).
β’ Experience with version control systems (Git Hub Actions) and collaborative development.
β’ Knowledge of cloud platforms, preferably AWS (S3, EC2, Lambda, etc.).
Preferred Qualifications:
β’ Master's degree in a relevant field.
β’ AWS certifications (Machine Learning Specialty, Solutions Architect, etc.).
β’ Knowledge of containerization and orchestration technologies.
β’ Experience with monitoring and observability tools (CloudWatch, Prometheus, etc.).
β’ Experience with big data technologies (EMR, Spark, Hadoop, etc.).
β’ Understanding of software engineering best practices and design patterns.
β’ Good working experience in ETL (SSIS or Sqoop/Spark).
β’ Experience with EMR
β’ Expert SQL knowledge (All types of Joins, CTEβs, Indexes, Stored Procedures, SQL performance).
β’ Knowledge in building basic machine learning models (Classification & Regression).
β’ Knowledge in Docker/MLOps and its orchestrations.
Key Skills & Competencies:
β’ Strong analytical and problem-solving abilities.
β’ Excellent communication and collaboration skills.
β’ Ability to work in fast-paced, agile environments.
β’ Detail-oriented with a focus on code quality and documentation.
β’ Continuous learning mindset and adaptability to new technologies.
β’ Experience working cross-functionally with data scientists, engineers, and product teams.





