Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is a Data Engineer contract for 6 months, fully remote but candidates must reside in GA. Requires 3+ years in data engineering, Python, SQL, and big data environments. Familiarity with GCP, ETL, and machine learning is essential.
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
April 22, 2025
🕒 - Project duration
More than 6 months
🏝️ - Location type
Remote
📄 - Contract type
W2 Contractor
🔒 - Security clearance
Unknown
📍 - Location detailed
Atlanta, GA
🧠 - Skills detailed
#AI (Artificial Intelligence) #Migration #Pig #Datasets #Big Data #GCP (Google Cloud Platform) #Scripting #Cloud #PySpark #SQL (Structured Query Language) #Hadoop #Programming #Libraries #ML (Machine Learning) #Spark (Apache Spark) #Data Science #Bash #BigQuery #Impala #Data Wrangling #"ETL (Extract #Transform #Load)" #Data Management #GitHub #Python #API (Application Programming Interface) #Consul #Scala #Informatica BDM (Big Data Management) #Automation #Data Engineering
Role description

Data Engineer

Exciting opportunity to achieve your true potential, chart new paths, develop new skills, collaborate with bright minds, and make a meaningful impact in an enterprise environment.

Candidates will collaborate with some of the best talents in the industry to create and implement innovative, high-quality solutions focused on our customers' needs.

RESPONSIBILITIES:

   • Work with data scientist team to migration the analytical data and projects to GCP environment and ensure the smooth project transition

   • Prepare and build data and analytical automation pipeline for self-serving machine learning projects: gather data from multiple sources and systems, integrating, consolidating and cleansing data, and structuring data for use by our clients our client facing projects.

   • Design and Code analysis scripts that can run on GCP using BigQuery/Python/Scala leverage multiple Core data sources

   • Build automated ML/AI modules, job, and data preparation pipelines by gathering data from multiple sources and systems, integrating, consolidating and cleansing data, and structuring data and analytical procedures for use by our clients in our solutions.

   • Perform design, creation, and interpretation of large and highly complex datasets

   • Consult with internal and external clients to understand the business requirements so successfully build datasets and implement complex big data solutions (under senior lead's supervision).

   • Ability to work with Technology and analytics teams to review, understand and interpret the business requirements to design and build missing functionalities to support the identity and fraud analytics needs (under senior lead's supervision).

   • Ability to work on the end to end interpretation , design, creation, and build of large and highly complex analytics related capabilities (under senior lead's supervision).

   • Strong oral and written communication skills, and ability to collaborate with cross-functional partners

This is a 6 month contract opportunity with a leading global organization based in Alpharetta, GA! This is a fully remote position - however resources need to reside in GA, local to Atlanta/Alpharetta and be willing to come onsite as needed/ in the beginning to ramp up. Successful candidates have a high level of initiative and thrive in a fast paced, enterprise environment.

Visionaire Partners offers all full-time W2 contractors a comprehensive benefits package for the contractor, their spouses/domestic partners, and dependents. Options include 401k with up to 4% match, medical, dental, vision, life insurance, short and long term disability, critical illness, hospital indemnity, accident coverage, and both Medical and Dependent Care Flexible Spending Accounts.

REQUIRED SKILLS:

   • 3+ years of professional data engineering or data wrangling experience

   • 3+ years working with Python and SQL

   • Experience with Hadoop based or Cloud based big data management environment

   • Bash scripting or similar experience for data movement and ETL

   • Big data queries in Hive/Impala/Pig/BigQuery (Sufficient in BigQuery API libraries to data prep automation is a huge plus)

   • Advanced Python programming including PySpark, with strong coding experience and Proficient in data studio, Big Table, GitHub working experience

   • Understand best practices for data management, maintenance, and reportingand use that knowledge to implement improvements

   • Basic knowledge in machine learning (ensemble machine learning models, unsupervised machine learning models)