Data Engineer

New Yesterday

This is a CONTRACT TO POSSIBLE HIRE POSITION!
We’re seeking a Mid-Level Data Engineer with hands-on experience in configuring and optimizing Apache Iceberg infrastructure. The role involves building out our foundational Iceberg data lakehouse architecture and integrating it with key cloud and analytics platforms. You will be a core part of our data engineering team, working closely with analytics and BI teams to ensure seamless data access and usability.
Key Responsibilities
Iceberg Infrastructure Configuration
Design and implement the initial Apache Iceberg infrastructure.
Ensure compatibility and optimization for batch and streaming data use cases.
Platform Integration & Connectivity
Set up and manage connections between Iceberg and:
Google Cloud Platform (GCP)
Snowflake – With a focus on Federated Data Warehousing (FDW) .
MicroStrategy
Looker
Power BI (lower priority, but still considered for downstream enablement)
Data Pipeline Development
Build and deploy initial pipelines for data flow from:
Snowflake → Iceberg → MicroStrategy
Monitor and optimize data ingestion, transformation, and delivery.
Ensure data quality, lineage, and security compliance throughout the pipeline.
Collaboration & Documentation
Collaborate cross-functionally with data science, analytics, and DevOps teams.
Document configuration, design patterns, and integration processes.
Qualifications: Qualifications
Required:
3–5 years of experience in data engineering or related field.
Proven experience configuring and managing Apache Iceberg environments.
Hands-on experience with Snowflake , including familiarity with FDW .
Experience integrating cloud storage systems and query engines (e.g., BigQuery, GCP).
Working knowledge of BI tools: MicroStrategy , Looker , Power BI .
Proficiency in Python, SQL, and data orchestration tools (e.g., Airflow).
Strong understanding of data lakehouse architecture and performance optimization.
Preferred:
Familiarity with secure data sharing and access control across tools.
Knowledge of metadata catalogs such as Apache Hive, AWS Glue, or Unity Catalog.
Background in working with distributed data systems and cloud-native environments.
Location:
Mettawa, IL, United States
Category:
Computer And Mathematical Occupations

We found some similar jobs based on your search