Vacancy expired!
- Experience in Java and Knowledge in Hadoop (HDFS/HBASE /SPARKSQL and SPARK/Scala with or prior experience in MapReduce) concepts and ability to write Spark/Scala RDD & SQL jobs
- Proven understanding with Hadoop, Spark, Hive, and Hbase and ability to write Shell scripting
- Familiarity with data loading tools like Sqoop, Flume and Kafka
- Knowledge of workflow/schedulers like Oozie
- Good aptitude in multi-threading and concurrency concept
- Loading data from disparate data source sets
- Certifications like Cloudera Developer (CCA175)/ Hortonworks Developer (Spark & Hive) /Administrator Certifications added advantage
- Hands on experience with at least two NO SQL databases
- Ability to analyze, identify issues with existing cluster, suggest architectural design changes
- Ability/Knowledge to implement Data Governance in Hadoop clusters
- Python
- KUDU
- Restful Services
- Machine Learning/Predictive analytical
- Spark streaming
- Ability to coordinate competing priorities and drive team work
- Ability to work in diverse team environments that are local and remote
- Strong Communication skills (Verbal and Written)
- Work with minimal supervision
- Collaborate with business analysts and line of business as needed
- Bachelor's degree from an accredited college or university
- 4-7 years' hands on experience in coding for the following Hadoop Eco System:
- Spark & Scala,
- Spark/RDD
- Spark-SQL
- HIVE
- Impala
- Java
- Oozie
- Autosys
- HBASE
- Shell Scripting
Vacancy expired!