

Lorven Technologies Inc.
AI/ML Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for an AI/ML Engineer with expertise in GraphDB, Neo4j, and entity resolution, based in Dallas, TX. It is a long-term contract position offering competitive pay. Key skills include anomaly detection, data enrichment, and model deployment.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
March 18, 2026
π - Duration
Unknown
-
ποΈ - Location
Hybrid
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Dallas, TX
-
π§ - Skills detailed
#AI (Artificial Intelligence) #Data Enrichment #Anomaly Detection #Data Governance #ML (Machine Learning) #Cloud #Neo4J #Knowledge Graph #Compliance #Deployment #Data Pipeline #Scala
Role description
Job Title: AI / ML Engineer with GraphDB / Knowledge Graph / Neo4j / Entity resolution.
Location: Dallas, TX (Hybrid)
Duration: Long Term contract
Job Description:
Client is looking for candidates who have experience in building:
β’ Ontology from large scale data (requires experience in entity resolution, probabilistic pattern matching)
β’ Agentic knowledge-based enrichment (automated data gap identification, and data enrichment)
β’ Anomaly detection on top of knowledge graph data at scale
β’ Fine tuning pipeline (including dataset generation, tuning, evaluation, deployment) for small language models and reasoning models
β’ Good depth in building models on top of unstructured data
About the Role:
β’ Clients is seeking a Sr Data and Gen AI Engineer with strong expertise in knowledge graph, Graph DB, Vector DB, Neo4j, RAG, and similar technologies.
β’ This engineer will design and implement data infrastructure that enables efficient fine-tuning and deployment of large language models (LLMs) on client servers for low-latency inference.
β’ The role demands a hands-on technologist who can architect, build, and optimize data systems serving enterprise-grade AI use cases.
Key Responsibilities:
β’ Design and implement GraphDB and VectorDB solutions to store, query, and retrieve structured and unstructured financial data.
β’ Build knowledge graph pipelines integrating multiple data sources to support LLM fine-tuning and retrieval-augmented generation workflows.
β’ Set up scalable data pipelines for model training, embedding generation, and data preprocessing
β’ Collaborate with AI researchers and ML engineers to prepare data and infrastructure for fine-tuning open-source or proprietary LLMs.
β’ Deploy and optimize model hosting for fast inference on on-prem or cloud GPU servers.
β’ Ensure data governance, lineage and compliance with internal and regulatory standards.
Looking forward for your reply.
Job Title: AI / ML Engineer with GraphDB / Knowledge Graph / Neo4j / Entity resolution.
Location: Dallas, TX (Hybrid)
Duration: Long Term contract
Job Description:
Client is looking for candidates who have experience in building:
β’ Ontology from large scale data (requires experience in entity resolution, probabilistic pattern matching)
β’ Agentic knowledge-based enrichment (automated data gap identification, and data enrichment)
β’ Anomaly detection on top of knowledge graph data at scale
β’ Fine tuning pipeline (including dataset generation, tuning, evaluation, deployment) for small language models and reasoning models
β’ Good depth in building models on top of unstructured data
About the Role:
β’ Clients is seeking a Sr Data and Gen AI Engineer with strong expertise in knowledge graph, Graph DB, Vector DB, Neo4j, RAG, and similar technologies.
β’ This engineer will design and implement data infrastructure that enables efficient fine-tuning and deployment of large language models (LLMs) on client servers for low-latency inference.
β’ The role demands a hands-on technologist who can architect, build, and optimize data systems serving enterprise-grade AI use cases.
Key Responsibilities:
β’ Design and implement GraphDB and VectorDB solutions to store, query, and retrieve structured and unstructured financial data.
β’ Build knowledge graph pipelines integrating multiple data sources to support LLM fine-tuning and retrieval-augmented generation workflows.
β’ Set up scalable data pipelines for model training, embedding generation, and data preprocessing
β’ Collaborate with AI researchers and ML engineers to prepare data and infrastructure for fine-tuning open-source or proprietary LLMs.
β’ Deploy and optimize model hosting for fast inference on on-prem or cloud GPU servers.
β’ Ensure data governance, lineage and compliance with internal and regulatory standards.
Looking forward for your reply.






