Vacancy expired!
- Work alongside Data Warehouse team (ETL Developers and Data Modelers/Architects/Engineers) recently migrated from on-prem to Azure Cloud, to design/develop/deploy new data warehouse (greenfield)
- Work with the business to define data needs
- Develop enterprise scale pipelines and data structures to ingest, organize, and regulate data for business user reporting needs
- Develop processes to successfully collect various data sources into Azure Data Lake - partition, curate, ingest, and integrate data
- Performance tuning - tune to best you can, then add horsepower as needed
- Automate data integration
- Large enterprise (mid to large organization) experience
- 1+ years Spark/Pyspark experience using Synapse (preferred) or Data Bricks
- 2+ years Data Warehouse/Engineering experience
- 1+ years Azure Cloud (must have Azure, not currently open to AWS or Google Cloud Platform currently) including Databricks, Synapse Analytics/Pipeline, Azure Data Lake/Factory, ADLS, Azure Event Hub, Azure Stream Analytics, Azure DevOps, Azure SQL DB, Dedicated SQL Pool, Apache Spark Pool, Azure Analysis Service, Serverless Pool, Azure Functions
- Cloud Optimization experience
- Understand how to physically change where data is stored
- ETL (Extract, Transform and Load)
- Python scripting
- Understand basic data modeling principals
- Streaming, batch (hybrid data processing) experience
- Technofunctional ability to handle both user/tech needs with a sense of urgency in an enterprise, multi-faceted environment
- CI/CD
Vacancy expired!