

Big Data Engineer (Locals & W2 Preferred)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Big Data Engineer with a contract length of "unknown," offering a pay rate of "unknown." Work is remote, with a preference for locals and W2 candidates. Key skills include Oracle, PL/SQL, Hadoop, Hive, and Python/PySpark.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
June 17, 2025
π - Project duration
Unknown
-
ποΈ - Location type
Unknown
-
π - Contract type
W2 Contractor
-
π - Security clearance
Unknown
-
π - Location detailed
Charlotte, NC
-
π§ - Skills detailed
#NumPy #Deployment #Scripting #Spark SQL #HDFS (Hadoop Distributed File System) #AI (Artificial Intelligence) #SciPy #Pandas #SQL (Structured Query Language) #Python #Spark (Apache Spark) #Oracle #Hadoop #Big Data #Data Ingestion #PySpark #Programming #Agile #"ETL (Extract #Transform #Load)" #Data Engineering #ML (Machine Learning) #Jupyter #RDBMS (Relational Database Management System)
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Local Prefers and Only W2 Or H1 Transfer.
Responsibilities include but are not limited to: Developing and enhancing application components for supporting ML/AI models and data ingestion processes, with focus on code resiliency and stability. Interacting & leading a team of developers, and interacting with business partners and develop processes to ensure that ML/AI models are production-ready Developing, enhancing, modifying and/or maintaining applications Working in a fast paced agile environment, under minimal supervision, with guidance from senior team members Participating in analysis on operational issues Participating in peer reviews for designs, code, and other work productsStrong knowledge of Oracle, SQL, RDBMS along with Python, Hadoop, Hive, Spark Experience in developing Hive & DBMS based applications Python programming background (scripting and object-oriented design) Coding experience with βbig dataβ (Spark/PySpark, SQL, Hadoop, ETL development) Experience implementing statistical models in python (Jupyter notebooks, scipy, numpy, pandas, Scikit-learn) Machine learning experience or knowledge
Overview:
Assess requirement and evaluate existing solutions
Build Process to interact with HDFS and Oracle using Python/ PySpark and Oracle PL/ SQL
Create Workflows, jobs and schedule them using Autosys
Works across development teams to contribute to the story refinement and delivery of data requirements through the delivery life cycle
Leverages architecture components in solution development, codes solutions to integrate, clean, transform, and control data as per acceptance criteria
Develops and executes test plans to produce quantitative results, identifies test issues and errors, and triages underlying causes
Drives complex information technology projects to ensure on-time delivery and adheres to team delivery and release processes
Identifies, defines, and documents data engineering requirements, communicating required information for deployment, maintenance, support, and business functionality
Ability to work independently with solid analytical skills
Ability to work with the team; excellent team player with great attitude
Data Resiliency Capabilities
Top 3 skills:
1. Oracle & PL/SQL Knowledge (Expert level)
1. Hadoop ecosystem, Hive Tables
1. Python/PySpark
Preferred Skills:
1. Autosys
1. Agile
Other Required Skills:Strong knowledge of Oracle, SQL, RDBMS along with Python, Hadoop, Hive, Spark Experience in developing Hive & DBMS based applications Python programming background (scripting and object-oriented design) Coding experience with βbig dataβ (Spark/PySpark, SQL, Hadoop, ETL development) Experience implementing statistical models in python (Jupyter notebooks, scipy, numpy, pandas, Scikit-learn) Machine learning experience or knowledge