Canvendor

Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer in San Jose, CA, on a contract basis for 10+ years of experience. Key skills include Power BI, ETL, Azure, and Python. Strong expertise in data pipelines, SQL optimization, and dimensional modeling is required.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
April 21, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
San Jose, CA
-
🧠 - Skills detailed
#Data Migration #Cloud #Jira #Pandas #Data Pipeline #Spark (Apache Spark) #NumPy #Azure Data Factory #DAX #Apache Spark #Storytelling #Python #GIT #ADF (Azure Data Factory) #Indexing #SSIS (SQL Server Integration Services) #"ETL (Extract #Transform #Load)" #Data Integration #Data Engineering #PySpark #Big Data #Migration #Version Control #Azure #SQL (Structured Query Language) #BI (Business Intelligence) #Data Manipulation #Scala #Microsoft Power BI
Role description
Role : Data Engineer Location : San Jose, CA ( Onsite) Type: Contract Experience – 10+ Years Key Skills : Power BI, ETL, Azure, Python JD : We are seeking a skilled data expert to transform complex, raw data into actionable business insights. This role focuses on designing, building, and maintaining scalable data pipelines using Azure Fabric and SSIS, ensuring our Sales (SFDC), Finance (NAV/BC), and Product (MSC) data is unified and ready for advanced reporting in Power BI. Skill Set & Expertise Requirements • Data Engineering: SSIS, Azure Data Factory, Azure Fabric (Pipelines, Medallion Architecture) • Design end-to-end Medallion Architecture, manage complex data migrations • Languages: Advanced SQL, Python (Pandas/NumPy) • Proficiency in performance tuning, indexing, and complex Python-based data manipulation • Big Data: Apache Spark, Azure Fabric Notebooks • Experience in distributed processing and PySpark-based notebook development. • Data Modelling: Kimball Methodology (Facts/Dimensions) • Deep understanding of dimensional modeling for BI performance. • BI & Reporting: Power BI, DAX, Power Query • Ability to create complex DAX measures and professional-grade storytelling dashboards. • Platforms: Gainsight, SFDC, NAV/BC, Jira • Hands-on experience integrating these diverse data sources into a central warehouse. • Tools: Git (Version Control), MSPO • Essential for collaborative development and environment management. Key Responsibilities Data Engineering & ETL • Pipeline Development: Design and deploy end-to-end ETL/ELT processes using Azure Data Factory and Azure Fabric (Notebooks & Pipelines). • Architecture Management: Implement and maintain Medallion Architecture (Bronze/Silver/Gold) and Kimball Methodology (Star Schemas) to ensure high-performing Data Warehousing. • Legacy Maintenance: Manage existing SSIS packages to ensure seamless data flow from on-prem and cloud sources. Reporting & Analytics • Dashboarding: Build sophisticated Power BI reports using Power Query for data transformation and DAX for complex calculations. • Platform Management: Administer the Power BI Service and manage data integration with Gainsight to drive Customer Success initiatives. • Source Integration: Extract and sync data from diverse sources including SFDC, NAV/BC, Jira, and product usage data (MSC). Database & Collaboration • SQL Optimization: Write and optimize complex Stored Procedures, Views, and Indexes for efficient data retrieval. • Version Control: Maintain codebase integrity using Git for version control and collaborative development.