

ALIS Software LLC
Knowledge Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Knowledge Engineer specializing in Enterprise AI & Agentic Systems on a long-term contract basis. Key skills include Databricks, advanced SQL, Python, and experience with RAG architecture. Remote work is offered, with a focus on data governance and AI readiness.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
March 5, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Remote
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
United States
-
🧠 - Skills detailed
#Data Governance #Automation #Data Engineering #Langchain #IAM (Identity and Access Management) #Security #Python #Data Enrichment #Indexing #Knowledge Graph #Data Modeling #Data Lineage #Metadata #Databricks #"ETL (Extract #Transform #Load)" #Azure #AI (Artificial Intelligence) #SQL (Structured Query Language) #Storage #Classification
Role description
Role: Knowledge Engineer – Enterprise AI & Agentic Systems
Remote
Type: Contract
Duration: Long Term
Role Summary
The Knowledge Engineer owns the transformation of raw enterprise data (structured and unstructured) into governed, reusable, AI-ready knowledge assets that power enterprise-grade GenAI and Agentic AI solutions. This role bridges data engineering, semantic modeling, data governance, and GenAI readiness, ensuring that enterprise agents are built on trusted, secure, governed knowledge foundations. The Knowledge Engineer thinks in terms of how AI consumes knowledge — optimizing for retrieval, reasoning, and agent orchestration, not just storage and transformation
Key Responsibilities
Enterprise Knowledge Foundation
• Convert structured sources (SQL, Delta tables) and unstructured repositories (NetDocuments, PDFs, contracts, emails) into clean, enriched, AI-consumable knowledge assets
• Design and implement semantic layers, metadata enrichment frameworks, ingestion pipelines, embeddings pipelines that serve LLM and Agentic AI consumption patterns
• Build reusable knowledge abstractions (entity models, ontologies, knowledge graphs) that scale across multiple AI use cases
Databricks-Centric Knowledge Engineering
• Build and manage end-to-end knowledge pipelines using Databricks
• Implement the full pipeline lifecycle: ingestion → transformation → chunking → embedding → indexing → retrieval
• Produce AI-consumption-ready knowledge layers with required quality guardrails
Governance, Catalog & Lineage
• Define & implement enterprise Data & AI governance for knowledge assets: data classification, RBAC/ABAC access controls, PII detection, tagging, masking, and lineage tracking
• Ensure all AI knowledge assets meet legal, regulatory, and internal security standards before reaching AI systems
Knowledge Architecture for Agentic AI
• Design knowledge structures optimized for: RAG pipelines, AI agents
• Implement retrieval optimization strategies including hybrid search (vector + keyword + metadata filtering)
• Build reusable entity relationships and context orchestration patterns that agents can reliably invoke at runtime
AI Platform Collaboration & Enablement
• Partner with AI Platform Engineers to expose governed, low-latency knowledge endpoints consumable by agent frameworks and MCP servers
• Contribute to prompt-context design by advising on knowledge structure, chunking strategies, and retrieval quality
• Act as SME on knowledge quality, helping AI teams debug retrieval failures and hallucination sources
Required Skills & Experience
Data & Knowledge Engineering -
• Strong hands-on experience with Databricks
• Advanced SQL and dimensional/semantic data modeling
• Python proficiency for pipeline development, transformation logic, and tooling automation
• Experience managing unstructured document repositories (like NetDocuments)
• Proficiency with vector database setup, configuration, and optimization (e.g., Pinecone, Weaviate, pgvector, Azure AI Search)
AI & Semantic Layer
• Deep knowledge of RAG architecture — including chunking strategies, embedding model selection, & retrieval evaluation
• Hands-on experience with LLM orchestration frameworks like LangChain, LlamaIndex, Microsoft Agent framework etc
• Familiarity with embedding optimization techniques and context management for production AI workloads
Governance & Security
• Experience implementing data lineage tracking
• Hands-on with data masking, classification pipelines, and PII handling at scale
• Enterprise IAM integration — applying RBAC/ABAC models to data and AI asset access
Role: Knowledge Engineer – Enterprise AI & Agentic Systems
Remote
Type: Contract
Duration: Long Term
Role Summary
The Knowledge Engineer owns the transformation of raw enterprise data (structured and unstructured) into governed, reusable, AI-ready knowledge assets that power enterprise-grade GenAI and Agentic AI solutions. This role bridges data engineering, semantic modeling, data governance, and GenAI readiness, ensuring that enterprise agents are built on trusted, secure, governed knowledge foundations. The Knowledge Engineer thinks in terms of how AI consumes knowledge — optimizing for retrieval, reasoning, and agent orchestration, not just storage and transformation
Key Responsibilities
Enterprise Knowledge Foundation
• Convert structured sources (SQL, Delta tables) and unstructured repositories (NetDocuments, PDFs, contracts, emails) into clean, enriched, AI-consumable knowledge assets
• Design and implement semantic layers, metadata enrichment frameworks, ingestion pipelines, embeddings pipelines that serve LLM and Agentic AI consumption patterns
• Build reusable knowledge abstractions (entity models, ontologies, knowledge graphs) that scale across multiple AI use cases
Databricks-Centric Knowledge Engineering
• Build and manage end-to-end knowledge pipelines using Databricks
• Implement the full pipeline lifecycle: ingestion → transformation → chunking → embedding → indexing → retrieval
• Produce AI-consumption-ready knowledge layers with required quality guardrails
Governance, Catalog & Lineage
• Define & implement enterprise Data & AI governance for knowledge assets: data classification, RBAC/ABAC access controls, PII detection, tagging, masking, and lineage tracking
• Ensure all AI knowledge assets meet legal, regulatory, and internal security standards before reaching AI systems
Knowledge Architecture for Agentic AI
• Design knowledge structures optimized for: RAG pipelines, AI agents
• Implement retrieval optimization strategies including hybrid search (vector + keyword + metadata filtering)
• Build reusable entity relationships and context orchestration patterns that agents can reliably invoke at runtime
AI Platform Collaboration & Enablement
• Partner with AI Platform Engineers to expose governed, low-latency knowledge endpoints consumable by agent frameworks and MCP servers
• Contribute to prompt-context design by advising on knowledge structure, chunking strategies, and retrieval quality
• Act as SME on knowledge quality, helping AI teams debug retrieval failures and hallucination sources
Required Skills & Experience
Data & Knowledge Engineering -
• Strong hands-on experience with Databricks
• Advanced SQL and dimensional/semantic data modeling
• Python proficiency for pipeline development, transformation logic, and tooling automation
• Experience managing unstructured document repositories (like NetDocuments)
• Proficiency with vector database setup, configuration, and optimization (e.g., Pinecone, Weaviate, pgvector, Azure AI Search)
AI & Semantic Layer
• Deep knowledge of RAG architecture — including chunking strategies, embedding model selection, & retrieval evaluation
• Hands-on experience with LLM orchestration frameworks like LangChain, LlamaIndex, Microsoft Agent framework etc
• Familiarity with embedding optimization techniques and context management for production AI workloads
Governance & Security
• Experience implementing data lineage tracking
• Hands-on with data masking, classification pipelines, and PII handling at scale
• Enterprise IAM integration — applying RBAC/ABAC models to data and AI asset access





