

Rconnect Consulting Inc
Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with a contract length of "unknown," offering a pay rate of "unknown." Key skills include Python, ETL pipelines, Azure SharePoint/M365 integration, OAuth/SSO flows, and custom API ingestion. Industry experience in data engineering is essential.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
March 18, 2026
π - Duration
Unknown
-
ποΈ - Location
Unknown
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
United States
-
π§ - Skills detailed
#Monitoring #Azure #JSON (JavaScript Object Notation) #Security #API (Application Programming Interface) #REST API #Indexing #Python #"ETL (Extract #Transform #Load)" #Data Engineering #Jira #Libraries #REST (Representational State Transfer) #Metadata #SharePoint #SaaS (Software as a Service)
Role description
JD:
Core engineering & data skills
β’ Strong in Python, ideally using official SDKs to call REST APIs.
β’ Experience building ETL / ingestion pipelines: incremental syncs, idempotent upserts, backfills, error handling, retries and rate-limit aware design.
β’ Comfort with JSON schema, pagination, and large-volume data movement.
Azure SharePoint / M365 integration
β’ Practical knowledge of SharePoint Online / OneDrive:
β’ Sites, document libraries, folders, item permissions, sensitivity labels, and how these affect what Glean can crawl and show.
β’ Experience integrating apps with Azure AD + M365:
β’ App registrations, Graph/SharePoint permissions, admin consent, and secure secret handling.
β’ Ability to debug Glean SharePoint/OneDrive connector behavior using admin console status (crawl progress, indexing errors, filters).
Authentication, security, and governance
β’ Hands-on with OAuth / SSO flows (esp. Entra ID / Azure AD) and configuring SaaS integrations:
β’ Understanding of bearer tokens, consent, scopes, redirect URIs.
β’ Experience with Glean-issued tokens:
β’ User vs global tokens, scopes (SEARCH, CHAT, DOCUMENTS, AGENTS, etc.), and X-Glean-ActAs for global tokens.
β’ Ability to map source-system ACLs to Glean permissions, ensuring permission-aware search results (SharePoint/OneDrive sites, Confluence spaces, Jira projects).
Custom ingestion using Gleanβs APIs
β’ Design and implementation of custom connectors where no native data source exists:
β’ Creating data sources, pushing documents + metadata + permissions via Gleanβs indexing API.
β’ Building robust ingestion jobs:
β’ Delta detection from upstream systems, resume/replay, dead-letter handling, and monitoring (logs/metrics) for ingest health.
JD:
Core engineering & data skills
β’ Strong in Python, ideally using official SDKs to call REST APIs.
β’ Experience building ETL / ingestion pipelines: incremental syncs, idempotent upserts, backfills, error handling, retries and rate-limit aware design.
β’ Comfort with JSON schema, pagination, and large-volume data movement.
Azure SharePoint / M365 integration
β’ Practical knowledge of SharePoint Online / OneDrive:
β’ Sites, document libraries, folders, item permissions, sensitivity labels, and how these affect what Glean can crawl and show.
β’ Experience integrating apps with Azure AD + M365:
β’ App registrations, Graph/SharePoint permissions, admin consent, and secure secret handling.
β’ Ability to debug Glean SharePoint/OneDrive connector behavior using admin console status (crawl progress, indexing errors, filters).
Authentication, security, and governance
β’ Hands-on with OAuth / SSO flows (esp. Entra ID / Azure AD) and configuring SaaS integrations:
β’ Understanding of bearer tokens, consent, scopes, redirect URIs.
β’ Experience with Glean-issued tokens:
β’ User vs global tokens, scopes (SEARCH, CHAT, DOCUMENTS, AGENTS, etc.), and X-Glean-ActAs for global tokens.
β’ Ability to map source-system ACLs to Glean permissions, ensuring permission-aware search results (SharePoint/OneDrive sites, Confluence spaces, Jira projects).
Custom ingestion using Gleanβs APIs
β’ Design and implementation of custom connectors where no native data source exists:
β’ Creating data sources, pushing documents + metadata + permissions via Gleanβs indexing API.
β’ Building robust ingestion jobs:
β’ Delta detection from upstream systems, resume/replay, dead-letter handling, and monitoring (logs/metrics) for ingest health.






