Vacancy expired!
- Gathering requirements, designing, implementing and testing applications.
- Should ready to work on documentation of existing modules
- To work on the enhancements in Hadoop (Spark / Hive/ Unix shell scripting) ecosystem’s
- Accessing large data from a variety of sources, including Hadoop, Spark, Hive, HBase and Oracle.
- Supporting data driven, rule-based processing systems that are low maintenance and follow software development best practices.
- Working in a highly collaborative, agile environment with minimal supervision.
- Managing multiple priorities and projects simultaneously.
- Finding solutions to problems and demonstrating the ability to set and meet reasonable deadlines.
- Writing documentation in clear language that can be understood by non-technical team members.
- Bachelor degree or above in Computer Science, Engineering or related fields. Masters or above preferred.
- 5+ years of experience with Hadoop eco-systems such as hive, spark.
- 5+ years of experience with UNIX platform and proficient in Shell programming.
- Experienced and In-depth programming skills in at least one of the following languages: Scala, Python, Java.
- 2+ years of experience with Cloud technologies. Prior experience with Azure is a big plus.
- 2+ years of experience designing and implementing ETL is a big plus.
- Experience with processing large data volume and data process performance optimization.
- Strength in data analysis and detailed oriented
- Ability to influence others and operate effectively in a matrix organization
- Ability to work both independently and within a team environment
- Strong written and verbal communication skills are required.
Vacancy expired!