Vacancy expired!
- At least 5+ years of experience on designing and developing Data Pipelines for Data Ingestion or Transformation using Java or Scala or Python
- Knowledge of Big Data & Spark with AWS / EMR or Data Proc background will be beneficial
- At least4 years of experience in the following Big Data frameworks: File Format (Parquet, AVRO, ORC), Resource Management, Distributed Processing and RDBMS
- At least 4 years of developing applications with Monitoring, Build Tools, Version Control, Unit Test, TDD, Change Management to support DevOps
- At least 2 years of experience with SQL and Shell Scripting experience
- Experience of designing, building, and deploying production-level data pipelines using tools from Hadoop stack (HDFS, Hive, Spark, HBase, Kafka, NiFi, Oozie, Apache Beam, Apache Airflow etc).
- Experience with Spark programming (pyspark or scala or java).
- At least4 years of experience in the following Big Data frameworks: File Format (Parquet, AVRO, ORC), Resource Management, Distributed Processing and RDBMS
- At least 4 years of developing applications with Monitoring, Build Tools, Version Control, Unit Test, TDD, Change Management to support DevOps
- Knowledge of Big Data & Spark with AWS / EMR or Data Proc background will be beneficial
Vacancy expired!