Software Engineer - Data Engineering & ML
1 Days Old
3 months ago
Full Time
Bay Area, CA
Washington
West Coast, USA
Hybrid - Open to remote
140K - 180K/year
Apply
Location:
Bay Area, CA or Washington; Bay Area Preferred, Remote option
Term:
Full-Time; Permanent
Rhizome is seeking a Software Engineer who can scale Data Engineering in support of Machine Learning Development at Utility Scale. The ideal candidate will have a strong background in Data Processing Pipelines, DAGs, ETLs, Feature extraction, and Statistical Analytics using Python and AWS cloud.The ideal candidate will have deep expertise in working with GIS data, Relational Databases, CSVs, and Excel at Utility scale. Successful candidates will also have practical experience building large scale ETL pipelines on AWS or GCP for Data Engineering, Feature Extraction, Statistical Analysis and Correlations.
About Rhizome
Rhizome is at the forefront of developing decision intelligence technology at the intersection of climate science and infrastructure systems. Our team pursues this endeavor with the wisdom and steadiness of industry veterans, and the curiosity, grit, and energy of startup and technology enthusiasts.
Our climate resilience SaaS platform helps utilities, governments, and industries plan for greater resilience to climate change and extreme weather by applying AI to a vast amount of information that characterizes infrastructure assets and their vulnerability to extreme weather. Focused on the $500B resilience investment gap in the grid today, our mandate is simple: Help electric utilities proactively adapt to climate change by integrating cutting-edge climate-asset intelligence into their existing planning workflows. As the world experiences record-breaking climate-related impacts, especially related to grid failures, our platform identifies future extreme weather vulnerabilities on utility assets at high resolutions and empowers planners to optimize investment deployments that keep society safe during natural hazard events.
Roles and Responsibilities
Design, construct, and maintain data pipelines to combine large volumes of geospatial, climate + weather, and electric utility datasets
Work with cross-functional team to deliver data in support of analytic and ML pipelines
Develop deep familiarity with electric utility datasets and take ownership of integration of new datasets into our existing environments
Contribute to ML model development in context of reliability and resiliency for the grid
Optimize storage and ETL pipelines
Develop versioned, scalable, repeatable and reliable pipelines for utility data that is in GIS and Tabular format to Delta Lake format
Scale & Automate data pipelines for statistical analysis for internal and external use-cases
Exceptional ability to diagnose data issues and discrepancies
Ability to modularize different stages of data ingestion and verification
Ability to write algorithms for data sanity checks and classification of different data elements
Ability to develop heuristics and suggestions for missing data items
Ability to validate and unit test data pipelines
Qualifications
Exceptional Python programming skills
Experience and expertise in working with Earth Observation and other geospatial data, at the gigabyte-to-terabyte scale
Strong programming skills with NumPy, SciPy, Xarrays
Strong programming skills with frameworks like Dagster or Airflow or Prefect
Strong programming skills with Databricks or Apache Spark or Amazon EMR or Cloudera
Deep expertise in storage optimization and partitioning on RDS, Postgres, PostGIS, Delta Lake
Hands on with GIS dataset and QGIS or ESRI
Hands on Experience with multi-dimensional Climate or Weather data
Familiarity or hands on experience with Secure Cloud Development
Exceptional ability to diagnose data issues and discrepancies
Ability to modularize different stages of data ingestion and verification
Ability to write algorithms for data sanity checks and classification of different data elements
Ability to develop heuristics and suggestions for missing data items
Ability to validate and test pipelines and write functional test to validate the pipelines
We’ll pay extra close attention if you have:
Exposure or experience with Electric Utility Tech Stack
Exposure to applied ML and Data Engineering with Electric Utility background
Experience in early-stage startup environments
Culture and Core Values
At Rhizome, we lead with compassion and empathy, aiming to understand before we help. Our thesis as technologists is that, in order to fulfill our mission to protect society from the impacts of climate change through intentional, intelligent infrastructure planning, we need to embark on a journey of respectfully listening, learning, and then problem-solving. This sentiment is represented through our core values:
Empathy: Understanding and relating to problems, customers, and each other, with humility.
Creativity: Exploring with curiosity and building with intention.
Aspiration: Striving for societal impact, personal fulfillment, and simply doing good work.
Tenacity: Pushing past barriers and the status quo with a sense of optimism and determination.
Service Excellence: Delivering high-quality outcomes for our customers, colleagues, and communities.
Compensation and Benefits
Rhizome offers competitive salaries and an excellent package of benefits and stock options. Compensation is based on a variety of factors including experience, role, and location.
Rhizome Data
A changing climate demands Resilience by Design
We like solving hard problems with creativity, tenacity, and empathy for our customers. At the same time, we believe that being better stewards in our community, building lasting relationships, and connecting dots is critical to affecting long-lasting change. AI is what we build, and resilience is what we serve. We've assembled a team that you can count on, because at the end of the day, if the grid can be 99.9% reliable, why can't we?
#J-18808-Ljbffr
- Location:
- United States
- Job Type:
- FullTime
- Category:
- IT & Technology