techstaff-in

Neo4j Engineer / Applied Data Scientist (Contract Position)

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Neo4j Engineer / Applied Data Scientist on a contract basis, starting immediately until December 31, 2025. It offers a remote work location with a focus on Neo4j expertise, API development, and data modeling. A degree in Computer Science or Data Science is required.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
November 11, 2025
πŸ•’ - Duration
More than 6 months
-
🏝️ - Location
Remote
-
πŸ“„ - Contract
Fixed Term
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
United States
-
🧠 - Skills detailed
#Jira #API (Application Programming Interface) #Data Science #Data Lineage #Scala #AI (Artificial Intelligence) #Data Governance #Data Modeling #Data Ingestion #Classification #Neo4J #Python #Schema Design #Computer Science #HBase #Automation #Observability
Role description
πŸ›οΈ Neo4j Engineer / Applied Data Scientist (Contract Position) Position Summary We are seeking two highly skilled and experienced Neo4j Engineers / Applied Data Scientists to significantly enhance our data intelligence infrastructure. The primary focus of these roles is to design, extend, and operationalize our graph-based systems to deliver cross-customer insights, identify emerging patterns, and power our suite of AI-driven decision-making and recommendation tools. These are critical contract positions commencing immediately, with a defined end date. πŸ“… Contract Duration and Opportunity β€’ Contract Length: Starting immediately, concluding December 31, 2025. β€’ Type: Remote (Open to US-based candidates). β€’ Location Preference: We are actively prioritizing candidates in the Los Angeles, CA area, as one of these contract roles offers the strong potential for conversion to a full-time, hybrid position beginning in January 2026. Core Responsibilities The successful candidate will be responsible for end-to-end management and optimization of our data assets within the graph environment: β€’ Graph Schema Development: Extend and rigorously optimize the Neo4j schema to support complex relational data modeling (e.g., tiers, metrics, event nodes, and internal classifications). β€’ Data Ingestion and Classification: Develop robust ingestion and classification pipelines that automatically tag, route, and standardize new data inputs into the graph. β€’ API Development and Integration: Create and maintain resilient RESTful APIs for analytics, reporting, and seamless integration with key enterprise systems (HubSpot, Jira, Slack). β€’ Automation Logic Implementation: Build, test, and deploy sophisticated business logic for automated alerts, proactive recommendations, and responsive system actions. β€’ Cross-Functional Alignment: Collaborate closely with AI and Product Development teams to ensure data models accurately reflect and support real-world use cases and product requirements. Illustrative Projects β€’ Implementing and optimizing the Neo4j schema to define relationships between Tiers, Query Cohorts, and key performance indicators. β€’ Developing and securing API endpoints (e.g., /tiers/:tier/metrics, /events/:id) with comprehensive Role-Based Access Control (RBAC). β€’ Integrating prompt classification logic (e.g., $\text{Level 1}/\text{Level 2}/\text{Level 3}$ classifiers) directly into the data ingestion workflow. β€’ Engineering automation hooks to trigger critical system alerts via communications platforms and assign follow-up tasks in the task management system. Required Qualifications β€’ Experience: Minimum of three (3) years of direct, hands-on experience with Neo4j, demonstrating expertise in Cypher query optimization, schema design, and performance tuning. β€’ Education: Bachelor’s or Master’s degree in Computer Science, Data Science, or a closely related technical discipline. β€’ Technical Proficiency: Strong command of data modeling principles, graph traversal algorithms, and the development of secure, scalable APIs. β€’ Technology Stack: Demonstrated experience with Postgres and proficiency in a primary backend language, specifically Python or Node.js. β€’ Workflow Tools: Familiarity and practical experience integrating with enterprise automation and workflow tools (e.g., Slack, Jira, HubSpot equivalents). Desired Qualifications (Bonus) β€’ Previous experience working with or integrating AI systems and knowledge representation frameworks. β€’ Knowledge of data lineage, observability principles, and automated data governance practices.