Vacancy expired!
- Design/develop/implement Big Data/Hadoop platforms related to data ingestion, storage, transformation and analytics.
- Load data from disparate data sets, load Hive/HBase and RDBMS tables.
- Import and export data using Sqoop from HDFS to RDBMS.
- Preprocess data using Hive and Pig.
- Develop shell/Scala/Python scripts to transform the data in HDFS.
- Perform analysis of vast data stores and uncover insights.
- Create scalable and high-performance web services for data tracking.
- Create custom analytic and data mining algorithms for data extraction.
- Experience in developing shell/python scripts to transform the data in HDFS.
- Assist in resolution of infrastructure issues
- Execute and troubleshoot Spark and Hive jobs including performance tuning.
- 4+ years' experience with Hive/HBase/MRV1/MRV2
- Experience with Hadoop, HDFS, Hive, Apache Spark, Storm, and Kafka
- Experience with Java, Scala, Python, Impala, R, Json, Knox, YARN, Tomcat, Ambari
- Experience with ETL tools such as Sqoop and Pig
- Experience with RDBMS, SQL, MongoDB, and hierarchical data management
Vacancy expired!