CaryTek

Data Analyst with AI/ML and Databricks

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Analyst with AI/ML and Databricks in New York City, offering $55/hour on a W2 for a hybrid contract. Requires 7+ years in data science, proficiency in Pyspark, Python, SQL, and Azure Databricks.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
440
-
🗓️ - Date
May 7, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Hybrid
-
📄 - Contract
W2 Contractor
-
🔒 - Security
Unknown
-
📍 - Location detailed
New York, NY
-
🧠 - Skills detailed
#Microsoft Power BI #Data Cleansing #Data Science #Clustering #Datasets #Cloud #NLP (Natural Language Processing) #PySpark #Automation #ML (Machine Learning) #BI (Business Intelligence) #Data Governance #AI (Artificial Intelligence) #A/B Testing #Computer Science #ML Ops (Machine Learning Operations) #Model Validation #Pandas #Predictive Modeling #Databases #Data Quality #Python #SQL (Structured Query Language) #Classification #Statistics #Conceptual Data Model #DAX #Libraries #Azure Databricks #Data Engineering #Data Analysis #Mathematics #Data Mining #Data Pipeline #Azure #Forecasting #Visualization #Databricks #NumPy #Spark (Apache Spark)
Role description
Position : Data Analyst with AI/ML and databricks Location : New York City (no relocation candidates) Work Visa : H4 EAD. Green Card EAD. GC, US Citizen Only Pay: $55 On W2 Work Mode : Hybrid 3 days onsite and 2 day Remote Interview : Onsite Interview (Mandatory ) JOB SUMMARY Designing and maintaining data • Azure Databricks required & ML/AI JOB DESCRIPTION • Bachelor’s or Master’s degree in Computer Science, Data Science, Machine Learning, Statistics, Mathematics, or related field. • Strong experience in machine learning algorithms, predictive modeling, and data mining. • Proficiency in Pyspark, Python pandas (required) for data science workloads. • Strong SQL (required) knowledge and experience with relational databases. • Minimum 3 years of experience with data visualization tools such as Power BI, Dax Queries, and best practices. • Experience with Azure Databricks, Google Cloud, and modern data science libraries (e.g., scikit-learn, pandas, NumPy). • Experience with GenAI and large language models. • Ability to interpret complex datasets and produce actionable insights. • Must know how to analyze the root cause of dashboard errors. • Have experience in ML Ops and have strong coding background. • Have experience with Natural Language Processing (NLP). • Knowledge or experience with A/B Testing. • Working knowledge of designing, training, and implementing machine learning models. • Familiarity with cloud-based infrastructure • Excellent communication and problem-solving skills. • 7 or more years of experience in data science and machine learning engineering. Additional Skills (Skills that are a plus, but not required) • Knowledge of statistical methods and experimental design. Responsibilities • Key Responsibilities Advanced Analytics & Machine Learning • Design, develop, and optimize machine learning models (forecasting, classification, clustering). • Apply data mining techniques to uncover patterns and insights in large datasets. • Perform feature engineering, model validation, and performance tuning. • Explore and deploy modern AI and ML approaches to enhance automation and analytics. Data Preparation & Quality • Prepare structured and unstructured data for modeling and advanced analysis. • Develop scripts and tools for data cleansing, validation, and enrichment. • Collaborate with Data Engineering to maintain efficient data pipelines. • Identify data quality issues and propose remediation. Analytics, Insights & Reporting • Conduct deep-dive analyses to identify trends and improvement opportunities. • Communicate complex findings in clear, concise ways to technical and non-technical stakeholders. • Support the development of dashboards, metrics, and analytical solutions. Cross-Team Collaboration • Work with architects, engineers, and analysts to define analytical requirements. • Contribute to conceptual data model design and workflow optimization. • Promote best practices in machine learning, analytics, and data governance.