Vacancy expired!
- Ability to design, build and unit test the applications in Pyspark.
- In-depth knowledge of Hadoop, Spark, and similar frameworks
- Ability to understand existing Talend graph & logic to convert into Spark/PySpark/ Spark SQL.
- Knowledge of Hive, HDFS File System, HDFS File Types, HDFS compression codec.
- Experience in processing large amounts of structured and unstructured data, including integrating data from multiple sources.
- Experience working with Bitbucket and CI-CD process would be an added advantage.
- PySpark, Talend
- SQL
Vacancy expired!