Vacancy expired!
- Operational management and architecture of Hadoop ecosystem, managing 100s 1000s of nodes globally
- Build out clusters in data centers around the world? Tuning multi-tenant Hadoop ecosystem for operational efficiency, balancing various workloads and optimizing Yarn and Impala accordingly
- Implement security, encryption, authentication, and authorization controls to adhere to corporate security policies
- Support Data Governance and data lineage on the cluster
- Enable High Availability and resiliency in the cluster, achieving 99.9999% uptime
- Understand network optimization and DR strategies
- Support and help to drive our hybrid cloud strategy, develop strategies for compute burst
- Work with data architects on the logical data models and physical database designs optimized for performance, availability and reliability
- Helping to tuning and optimization of backend and frontend data operations
- Serve as a query tuning and optimization technical expert, providing feedback to team
- Scripting and automation to support development, QA and production database environments, deployments to production and management of services and infrastructure
- Mentors development team members
- Proactively helps to resolve difficult technical issues
- Provide technical knowledge to teams during project discovery and architecture phases
- Keep management informed of work activities and schedules
- Assess new initiatives to determine the work effort and estimate the necessary time-to- completion
- Document new development, procedures or test plans as needed
- Participate in data builds and deployment efforts. Help mature our Continuous Integration and Continuous Deployment methodologies
- Participate in projects through various phases
- Performs other related duties as assigned
- Partner with the business units to develop effective solutions that solve business challenges
- Cloudera CDP
- CM API
- Auto TLS - trusted certs
- Encryption at Rest
- Horton-Works HDP
- Experience with Atlas & Ranger policies setup
- Installation and configuration of the Ranger-KMS
- Migration & Upgrades from HDP to CDP
- Hadoop, Hive, Impala, HBase and related technologies
- Spark/Spark2
- MPP, shared nothing database systems, NoSQL systems
- Object Oriented and Functional Programming Experience
- Excellent knowledge of Linux, AIX, or other Unix flavors
- Experience with scripting (Bash, Linux scripting)
- Data Warehousing design and concepts
- Exposure to Infrastructure as Code (Ansible, Terraform)
Vacancy expired!