Flink with Datastreams API

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a "Flink with Datastreams API" freelancer, requiring 10 years in back-end Java development and 5 years with Kafka, Flink, SQL, and MongoDB. Contract length and pay rate are unspecified; remote work is available.
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
April 25, 2025
🕒 - Project duration
Unknown
🏝️ - Location type
Unknown
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
Dallas, TX
🧠 - Skills detailed
#Data Lake #API (Application Programming Interface) #SQL (Structured Query Language) #Azure ADLS (Azure Data Lake Storage) #Dynatrace #Azure #Java #MongoDB #Kafka (Apache Kafka) #GitHub #Cloud #Storage #ADLS (Azure Data Lake Storage) #Integration Testing
Role description

   • 10 years of hands-on design and java coding experience in back-end system development.

   • 5 years hands-on experience with Kafka, Flink, Cloud, Unit/Functional/Integration testing, SQL or kSQL, Java, Github Actions, Dynatrace, Code scanner, and MongoDB.

   • Expertise in Flink APIs (DataStream, Process functions, etc.).

   • Competence in state management (checkpoints and savepoints) with local storage.

   • Configuration of connectors like EventHub, Kafka, and MongoDB.

   • Implementation of Flink API Aggregators.

   • Handling watermarks for out-of-order events.

   • Management of state using Azure Data Lake Storage (ADLS).