

Cypress HCM
Senior GenAI Data Scientist
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior GenAI Data Scientist, offering a contract of unspecified length at a pay rate of $116.18 – 124.14/hr. Key skills include GenAI proficiency, advanced SQL, Python, and experience in sales analytics and data engineering.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
992
-
🗓️ - Date
February 10, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
W2 Contractor
-
🔒 - Security
Unknown
-
📍 - Location detailed
San Francisco Bay Area
-
🧠 - Skills detailed
#AI (Artificial Intelligence) #SQL (Structured Query Language) #Libraries #Data Pipeline #Compliance #Python #Snowflake #Documentation #Data Warehouse #AWS (Amazon Web Services) #Databases #Data Engineering #Airflow #GDPR (General Data Protection Regulation) #GitHub #"ETL (Extract #Transform #Load)" #Observability #Security #SQL Queries #ML (Machine Learning) #Data Quality #Data Science #CRM (Customer Relationship Management) #Langchain #NumPy #dbt (data build tool) #Cloud #Debugging #Pandas #GCP (Google Cloud Platform) #Apache Airflow #Azure #Data Integration #Slowly Changing Dimensions #BI (Business Intelligence)
Role description
Description
• We’re seeking an experienced contractor to architect, build, and productionize GenAI data science workflows that transform enterprise data into actionable business intelligence. This role sits at the intersection of generative AI, data engineering, and business analytics, requiring both deep technical expertise and the ability to collaborate effectively with business stakeholders.
• You’ll be working primarily on GenAI applications for sales intelligence, leveraging call transcripts and business data to deliver high-impact use cases in production.
Duties
• GenAI Engineering & Production
• Design and implement end-to-end GenAI workflows that integrate enterprise data sources (accounting/finance systems, sales call transcripts, CRM data)
• Build and deploy agentic AI workflows using frameworks like LangGraph, LangChain, or similar orchestration tools
• Implement comprehensive observability, evaluation frameworks, and guardrails for production GenAI systems
• Establish best practices for prompt engineering, retrieval-augmented generation (RAG), and model selection
• Critically evaluate use cases to determine when GenAI is (and isn’t) the appropriate solution
• Data Engineering & Architecture
• Design and implement data models including star schemas, slowly changing dimensions (SCD Type 2), and fact/dimension tables
• Write complex SQL queries to extract, transform, and analyze data from enterprise databases and data warehouses
• Build robust data pipelines using Apache Airflow for workflow orchestration
• Process and transform data using Python, pandas, and numpy
• Ensure data quality, governance, and compliance standards are met
• Business Intelligence & Stakeholder Management
• Translate business requirements into technical solutions, particularly around sales metrics, KPIs, and performance analytics
• Work directly with business stakeholders to gather requirements, manage expectations, and communicate timelines
• Provide strategic guidance on what’s feasible, what’s valuable, and what trade-offs exist
• Deliver clear documentation and presentations for both technical and non-technical audiences
Requirements
• Technical Expertise
• GenAI Proficiency: Deep hands-on experience with LLM applications, including observability tools, evaluation frameworks, and safety guardrails
• Agentic AI: Demonstrated experience building multi-agent or agentic workflows using LangGraph or similar frameworks
• LLM Fundamentals: Strong understanding of how LLMs work, their capabilities and limitations, context windows, tokenization, embeddings, and fine-tuning
• AI-Assisted Development: Active user of GenAI coding tools (Cursor, GitHub Copilot, Codex, Gemini Code Assist, etc.) with proven ability to accelerate development
• SQL Mastery: Expert-level SQL skills including complex joins, window functions, CTEs, query optimization, and performance tuning
• Data Engineering: Expert knowledge of dimensional modeling (star schemas, SCD Type 2), data warehouse concepts, and ETL/ELT patterns
• Python Stack: Advanced proficiency in Python, pandas, numpy, and related data science libraries
• Workflow Orchestration: Production experience with Apache Airflow or similar orchestration platforms
• Enterprise Data Integration: Experience working with structured data from ERP, CRM, and financial systems
• Business & Domain Knowledge
• Strong grasp of sales operations, pipeline metrics, conversion funnels, and revenue analytics
• Understanding of key business metrics and KPIs across sales, finance, and operations
• Ability to identify high-value use cases and prioritize based on business impact
• Experience analyzing sales conversations and extracting actionable insights
• Soft Skills & Work Style
• Communication: Excellent written and verbal communication skills; ability to explain complex technical concepts to non-technical stakeholders
• Stakeholder Management: Proven track record managing business partner relationships, setting realistic expectations, and delivering on commitments
• Independence: Self-directed and able to own projects end-to-end with minimal supervision
• Pragmatism: Bias toward shipping working solutions; comfortable with iteration and incremental delivery
• Problem Solving: Strong analytical and debugging skills; resourceful when facing ambiguous challenges
• Nice to Have:
• Experience with vector databases
• Knowledge of cloud platforms (AWS, GCP, Azure) and their AI/ML services
• Experience with dbt (data build tool) for analytics engineering
• Experience with streaming data and real-time processing
• Background in conversation intelligence or speech-to-text applications
• Understanding of privacy, security, and compliance requirements for AI systems (SOC 2, GDPR, etc.)
• Previous experience in a startup or fast-paced environment
• Familiarity with modern data warehouse solutions (Snowflake, Hive)
Compensation
• $116.18 – 124.14/hr W—2
Description
• We’re seeking an experienced contractor to architect, build, and productionize GenAI data science workflows that transform enterprise data into actionable business intelligence. This role sits at the intersection of generative AI, data engineering, and business analytics, requiring both deep technical expertise and the ability to collaborate effectively with business stakeholders.
• You’ll be working primarily on GenAI applications for sales intelligence, leveraging call transcripts and business data to deliver high-impact use cases in production.
Duties
• GenAI Engineering & Production
• Design and implement end-to-end GenAI workflows that integrate enterprise data sources (accounting/finance systems, sales call transcripts, CRM data)
• Build and deploy agentic AI workflows using frameworks like LangGraph, LangChain, or similar orchestration tools
• Implement comprehensive observability, evaluation frameworks, and guardrails for production GenAI systems
• Establish best practices for prompt engineering, retrieval-augmented generation (RAG), and model selection
• Critically evaluate use cases to determine when GenAI is (and isn’t) the appropriate solution
• Data Engineering & Architecture
• Design and implement data models including star schemas, slowly changing dimensions (SCD Type 2), and fact/dimension tables
• Write complex SQL queries to extract, transform, and analyze data from enterprise databases and data warehouses
• Build robust data pipelines using Apache Airflow for workflow orchestration
• Process and transform data using Python, pandas, and numpy
• Ensure data quality, governance, and compliance standards are met
• Business Intelligence & Stakeholder Management
• Translate business requirements into technical solutions, particularly around sales metrics, KPIs, and performance analytics
• Work directly with business stakeholders to gather requirements, manage expectations, and communicate timelines
• Provide strategic guidance on what’s feasible, what’s valuable, and what trade-offs exist
• Deliver clear documentation and presentations for both technical and non-technical audiences
Requirements
• Technical Expertise
• GenAI Proficiency: Deep hands-on experience with LLM applications, including observability tools, evaluation frameworks, and safety guardrails
• Agentic AI: Demonstrated experience building multi-agent or agentic workflows using LangGraph or similar frameworks
• LLM Fundamentals: Strong understanding of how LLMs work, their capabilities and limitations, context windows, tokenization, embeddings, and fine-tuning
• AI-Assisted Development: Active user of GenAI coding tools (Cursor, GitHub Copilot, Codex, Gemini Code Assist, etc.) with proven ability to accelerate development
• SQL Mastery: Expert-level SQL skills including complex joins, window functions, CTEs, query optimization, and performance tuning
• Data Engineering: Expert knowledge of dimensional modeling (star schemas, SCD Type 2), data warehouse concepts, and ETL/ELT patterns
• Python Stack: Advanced proficiency in Python, pandas, numpy, and related data science libraries
• Workflow Orchestration: Production experience with Apache Airflow or similar orchestration platforms
• Enterprise Data Integration: Experience working with structured data from ERP, CRM, and financial systems
• Business & Domain Knowledge
• Strong grasp of sales operations, pipeline metrics, conversion funnels, and revenue analytics
• Understanding of key business metrics and KPIs across sales, finance, and operations
• Ability to identify high-value use cases and prioritize based on business impact
• Experience analyzing sales conversations and extracting actionable insights
• Soft Skills & Work Style
• Communication: Excellent written and verbal communication skills; ability to explain complex technical concepts to non-technical stakeholders
• Stakeholder Management: Proven track record managing business partner relationships, setting realistic expectations, and delivering on commitments
• Independence: Self-directed and able to own projects end-to-end with minimal supervision
• Pragmatism: Bias toward shipping working solutions; comfortable with iteration and incremental delivery
• Problem Solving: Strong analytical and debugging skills; resourceful when facing ambiguous challenges
• Nice to Have:
• Experience with vector databases
• Knowledge of cloud platforms (AWS, GCP, Azure) and their AI/ML services
• Experience with dbt (data build tool) for analytics engineering
• Experience with streaming data and real-time processing
• Background in conversation intelligence or speech-to-text applications
• Understanding of privacy, security, and compliance requirements for AI systems (SOC 2, GDPR, etc.)
• Previous experience in a startup or fast-paced environment
• Familiarity with modern data warehouse solutions (Snowflake, Hive)
Compensation
• $116.18 – 124.14/hr W—2






