Vacancy expired!
- Design, build and maintain Big Data workflows/pipelines to process billions of records in large-scale data environments with experience in end-to-end design and build process of Near-Real Time and Batch Data Pipelines.
- Leads code review sessions to validate adherence with development standards and benchmark application performance by capacity testing.
- Experience with software testing frameworks.
- Leverages DevOps techniques and Experience with DevOps tools - GitHub, Jira, Jenkins, Crucible for Continuous Integration, Continuous Deployment and build automation.
- Develop, implement and optimize streaming, data lake, and big data analytics solutions
- Support reusable framework and data governance processes by partnering with LOBs for any code/requirements remediation
- Engage in application design and data modeling discussions also participate in developing and enforcing data security policies
- Knowledge of NoSQL, RDBMS, SQL, JSON, XML and ETL skills are must.
- Understanding of data transformations, cleansing, and deduplications.
- Advanced knowledge of SQL (PSQL or TSQL).
- Experience developing pipelines for both Cloud and Hybrid Cloud infrastructures.
- Experience in AWS utilizing services such as S3, AWS CLI, and RDS.
- Experience using modern ETL tools like Talend and Nifi.
- Experience working in an Agile delivery environment
- Ability to work independently and drive solutions end to end leveraging various technologies to solve data problems and develop solutions.
- Perform unit tests and conduct reviews with other team members to ensure code is designed with high code coverage
- Passionate for continuous learning, experimenting, applying and contributing towards cutting edge open source technologies and software paradigms
- Ability to research and assess open source technologies and components to recommend and integrate into the design and implementation
Vacancy expired!