

Data Engineer
Data Engineer
Exciting opportunity to achieve your true potential, chart new paths, develop new skills, collaborate with bright minds, and make a meaningful impact in an enterprise environment.
Candidates will collaborate with some of the best talents in the industry to create and implement innovative, high-quality solutions focused on our customers' needs.
RESPONSIBILITIES:
• Work with data scientist team to migration the analytical data and projects to GCP environment and ensure the smooth project transition
• Prepare and build data and analytical automation pipeline for self-serving machine learning projects: gather data from multiple sources and systems, integrating, consolidating and cleansing data, and structuring data for use by our clients our client facing projects.
• Design and Code analysis scripts that can run on GCP using BigQuery/Python/Scala leverage multiple Core data sources
• Build automated ML/AI modules, job, and data preparation pipelines by gathering data from multiple sources and systems, integrating, consolidating and cleansing data, and structuring data and analytical procedures for use by our clients in our solutions.
• Perform design, creation, and interpretation of large and highly complex datasets
• Consult with internal and external clients to understand the business requirements so successfully build datasets and implement complex big data solutions (under senior lead's supervision).
• Ability to work with Technology and analytics teams to review, understand and interpret the business requirements to design and build missing functionalities to support the identity and fraud analytics needs (under senior lead's supervision).
• Ability to work on the end to end interpretation , design, creation, and build of large and highly complex analytics related capabilities (under senior lead's supervision).
• Strong oral and written communication skills, and ability to collaborate with cross-functional partners
This is a 6 month contract opportunity with a leading global organization based in Alpharetta, GA! This is a fully remote position - however resources need to reside in GA, local to Atlanta/Alpharetta and be willing to come onsite as needed/ in the beginning to ramp up. Successful candidates have a high level of initiative and thrive in a fast paced, enterprise environment.
Visionaire Partners offers all full-time W2 contractors a comprehensive benefits package for the contractor, their spouses/domestic partners, and dependents. Options include 401k with up to 4% match, medical, dental, vision, life insurance, short and long term disability, critical illness, hospital indemnity, accident coverage, and both Medical and Dependent Care Flexible Spending Accounts.
REQUIRED SKILLS:
• 3+ years of professional data engineering or data wrangling experience
• 3+ years working with Python and SQL
• Experience with Hadoop based or Cloud based big data management environment
• Bash scripting or similar experience for data movement and ETL
• Big data queries in Hive/Impala/Pig/BigQuery (Sufficient in BigQuery API libraries to data prep automation is a huge plus)
• Advanced Python programming including PySpark, with strong coding experience and Proficient in data studio, Big Table, GitHub working experience
• Understand best practices for data management, maintenance, and reportingand use that knowledge to implement improvements
• Basic knowledge in machine learning (ensemble machine learning models, unsupervised machine learning models)
Data Engineer
Exciting opportunity to achieve your true potential, chart new paths, develop new skills, collaborate with bright minds, and make a meaningful impact in an enterprise environment.
Candidates will collaborate with some of the best talents in the industry to create and implement innovative, high-quality solutions focused on our customers' needs.
RESPONSIBILITIES:
• Work with data scientist team to migration the analytical data and projects to GCP environment and ensure the smooth project transition
• Prepare and build data and analytical automation pipeline for self-serving machine learning projects: gather data from multiple sources and systems, integrating, consolidating and cleansing data, and structuring data for use by our clients our client facing projects.
• Design and Code analysis scripts that can run on GCP using BigQuery/Python/Scala leverage multiple Core data sources
• Build automated ML/AI modules, job, and data preparation pipelines by gathering data from multiple sources and systems, integrating, consolidating and cleansing data, and structuring data and analytical procedures for use by our clients in our solutions.
• Perform design, creation, and interpretation of large and highly complex datasets
• Consult with internal and external clients to understand the business requirements so successfully build datasets and implement complex big data solutions (under senior lead's supervision).
• Ability to work with Technology and analytics teams to review, understand and interpret the business requirements to design and build missing functionalities to support the identity and fraud analytics needs (under senior lead's supervision).
• Ability to work on the end to end interpretation , design, creation, and build of large and highly complex analytics related capabilities (under senior lead's supervision).
• Strong oral and written communication skills, and ability to collaborate with cross-functional partners
This is a 6 month contract opportunity with a leading global organization based in Alpharetta, GA! This is a fully remote position - however resources need to reside in GA, local to Atlanta/Alpharetta and be willing to come onsite as needed/ in the beginning to ramp up. Successful candidates have a high level of initiative and thrive in a fast paced, enterprise environment.
Visionaire Partners offers all full-time W2 contractors a comprehensive benefits package for the contractor, their spouses/domestic partners, and dependents. Options include 401k with up to 4% match, medical, dental, vision, life insurance, short and long term disability, critical illness, hospital indemnity, accident coverage, and both Medical and Dependent Care Flexible Spending Accounts.
REQUIRED SKILLS:
• 3+ years of professional data engineering or data wrangling experience
• 3+ years working with Python and SQL
• Experience with Hadoop based or Cloud based big data management environment
• Bash scripting or similar experience for data movement and ETL
• Big data queries in Hive/Impala/Pig/BigQuery (Sufficient in BigQuery API libraries to data prep automation is a huge plus)
• Advanced Python programming including PySpark, with strong coding experience and Proficient in data studio, Big Table, GitHub working experience
• Understand best practices for data management, maintenance, and reportingand use that knowledge to implement improvements
• Basic knowledge in machine learning (ensemble machine learning models, unsupervised machine learning models)