Job Details

ID #41106354
State California
City Oakland
Job type Permanent
Salary USD TBD TBD
Source Blue Shield Of California
Showed 2022-05-18
Date 2022-05-17
Deadline 2022-07-16
Category Architect/engineer/CAD
Create resume

Data Integration Developer, Principal

California, Oakland, 94601 Oakland USA

Vacancy expired!

Blue Shield of California's mission is to ensure all Californians have access to high-quality health care at a sustainably affordable price. We are transforming health care in a way that truly serves our nonprofit mission by lowering costs, improving quality, and enhancing the member and physician experience.To fulfill our mission, we must ensure a diverse, equitable, and inclusive environment where all employees can be their authentic selves and fully contribute to meet the needs of the multifaceted communities we serve. Our comprehensive approach to diversity, equity, and inclusion combines a focus on our people, processes, and systems with a deep commitment to promoting social justice and health equity through our products, business practices, and presence as a corporate citizen.

Blue Shield has received awards and recognition for being a certified Great Place to Work, best place to work for LGBTQ equality, leading disability employer, one of the best companies for women to advance, Bay Area's top companies in volunteering & giving, and one of the world's most ethical companies. Here at Blue Shield of California, we are striving to make a positive change across our industry and the communities we live in - join us!

Your Role The Data Engineer builds, manages and optimizes production data pipelines supporting key data and analytics initiatives.You will be the cornerstone of a robust data stack, integrating diverse sources and serving up data to power machine learning solutions and self-service analytics. You will mine the operational details of adjudicating claims and build a real time pipeline to feed an AI-enabled recommendation engine. You will follow our members' journey through the healthcare landscape and blend data points across a broad spectrum of applications to create multidimensional records serving interventions. You will scrape surveys and munge data streams to extract measurable insights. From the minutiae of desk level procedures to macro population trends, there are no problems too big or small for your keen eyes and open mind.

Your Work In this role, you will:
  • Build, develop, implement and execute extensible reusable data pipelines consisting of multiple acquisition sources and integration into use case driven endpoints
  • Maintain and optimize workloads in various deployment stages and data environments to ensure optimal performance as data volume and variety increase
  • Lead design activities in partnership with data scientists, analysts and product owners to translate functional requirements into technical specifications for scalable data pipelines
  • Oversee management of analytical data assets for exploratory and early stage analytic usage patterns, and develop recommendations to integrate with production pipelines
  • Orchestrate data pipelines using modern tools and techniques to automate repeatable ETL processes, minimize error prone dependencies and improve integrity of published data assets
  • Collaborate with internal IT teams to troubleshoot incidents and coordinate resolutions to minimize disruption of analytic applications
  • Monitor data consumption patterns and develop enhancements to ensure pipelines adapt to evolving data schema and analytic use cases
  • Collaborate with data consumers to define and catalog use cases to ensure adherence to data governance standards and ethical/legal guidelines

Your Knowledge and Experience
  • A college degree or equivalent in computer science, data management, information systems or related quantitative field
  • A minimum of 10+ years of experience in data management disciplines. Industry experience in health care sector preferred.
  • High proficiency working with large, heterogeneous datasets in building/optimizing data pipelines using ELT, data replication, API access, data virtualization, stream data integration, and emerging technologie High proficiency with relational databases (Netezza, Oracle, MS SQL), NoSQL databases (MongoDB, Cassandra), and distributed computing platforms
  • Proficiency in Python, R, Scala, Julia or equivalent scripting language for data analysis
  • High proficiency with CI/CD tools and rigorous application of DataOps principles
  • Successful track record implementing complex automated data pipelines with commercial data preparation tools (Trifacta) or open-source technologies (Airflow, Spark)
  • Demonstrated ability to work across multiple deployment environments, operating systems and through containerization techniques such as Docker and Kubernetes
  • Ability to partner, collaborate with, and influence relevant stakeholders across diverse functions and experience levels
  • Strong independent judgment, critical thinking and problem-solving skills required to anticipate and respond to emerging challenges

Our Values
  • Honest. We hold ourselves to the highest ethical and integrity standards. We build trust by doing what we say we're going to do and by acknowledging and correcting where we fall short
  • Human. We strive to be our authentic selves, listening and communicating effectively, and showing empathy towards others by walking in their shoes
  • Courageous. We stand up for what we believe in and are committed to the hard work necessary to achieve our ambitious goals

#dice

Vacancy expired!

Subscribe Report job