

Data Scientist Premium
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Scientist Premium, requiring 4 days on-site. The contract pays "USD" and demands a Bachelor's or Master's in computer science with 3+ years of experience, proficiency in Python, SQL, and GCP, and expertise in data engineering and NLP.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
August 13, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Washington, DC
-
π§ - Skills detailed
#Data Pipeline #NLP (Natural Language Processing) #Cloud #Data Processing #ML (Machine Learning) #Documentation #"ETL (Extract #Transform #Load)" #AI (Artificial Intelligence) #Data Enrichment #SharePoint #Data Science #Python #Data Engineering #Computer Science #Databricks #Scala #SQL (Structured Query Language) #GCP (Google Cloud Platform) #Metadata
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Data Scientist Premium I [USD]
4 DAYS ON SITE
Data Source Analysis and Integration
β’ Analyze existing data sources, including SharePoint Online (documents) and Databricks SQL Warehouse (structured data).
β’ Design and implement methods for extracting and integrating data from these sources.
b. Data Enrichment Layer Design
β’ Develop techniques for text extraction, metadata standardization, and semantic enrichment.
β’ Design a scalable enrichment layer to process and enhance data for downstream AI/ML use.
c. Embedding Data Pipeline Development
β’ Build and automate ETL/ELT pipelines for embedding data in Vertex AI.
β’ Apply NLP and data processing techniques to ensure data is suitable for AI/ML models.
d. Documentation and Collaboration
β’ Prepare comprehensive technical documentation for all developed processes and pipelines.
β’ Collaborate with AI/ML teams to ensure enriched data meets project requirements.
1. Skills/Experience Required
β’ Bachelor's degree in computer science, data engineering, or related field with 3+ years of experience or Master's Degree in computer science, data engineering, or related field.
β’ Proven experience in data engineering and enrichment, especially with unstructured and structured data.
β’ Proficiency in Python, SQL, and cloud platforms (preferably GCP and Vertex AI).
β’ Experience with NLP techniques and data processing frameworks.
β’ Strong problem-solving, communication, and documentation skills.