Job Details

ID #21473782
State Texas
City Austin
Job type Contract
Salary USD $140,000+ 140000+
Source Consulting Solutions LLC
Showed 2021-10-21
Date 2021-10-08
Deadline 2021-12-06
Category Et cetera
Create resume

Hadoop Engineer - REMOTE

Texas, Austin, 60290 Austin USA

Vacancy expired!

Hadoop Engineer REMOTE

Top Skills:
  • Hadoop
o Hive o Sqoop o Oozie o Kafka o Spark Python o Scala
  • Spark
  • Cloudera

Soft Skills:
  • Support mindset

Day to Day Responsibilities:
  • Analysis, designing, development and support in Hadoop infrastructure using Hive, Sqoop, Oozie, Kafka, Spark Python and Scala.
  • Experience in working with Cloudera distribution with various nodes cluster running Spark on YARN.
  • Hands on experience in using Hive to extract, transform and load (ETL) data into a reportable format using Spark environment.
  • Experience in importing and exporting Gigabytes of data between HDFS and Relational Teradata Database using Sqoop.
  • HBase Database skills.
  • Experience in using Parquet file format in Hive and Spark.
  • Clear understanding of Hadoop architecture and various component like Resource manager, node manager, Name node, Data node HDFS etc.
  • Programing knowledge to get the data from Kafka using PySpark and load into Hadoop distributed file system and other downstream applications.
  • Hands on experience in designing ETL Informatica power center 9.1. Experience of Data warehousing concepts like Dimensional modeling like SCD-1, SCD-2 etc., different schemas (Star, Snowflakes etc.), Demoralization etc.
  • Proficient in Data analysis, Data modeling (Physical and logical), That Include ER diagrams, DFD (data flow diagrams) and Database designing.
  • Having Experience to work on Production support project consist of incident investigations, tracking the regular data loading jobs, resolve the tickets etc.
  • Proficient in performance tuning using through execution plan, Hints, Bulk Binding, Pipelined Functions, Partitions, index etc.
  • Having experience of UNIX Shell programing.
  • Pyspark and Scala programing experience.
  • Experience of Core Java, Angular J, JASON, Tableau, CA7, ServiceNow is preferred but not necessary.
  • Day to day development, support and monitoring of our production environments.
  • Leading troubleshooting calls and driving to root cause resolution; makes recommendations for long-term resolution of problems across the enterprise.
  • Provide systems support in and out of office hours.
  • Assist our business partners with customer issues when needed
  • Measure and optimize application performance. Configure, tune and troubleshoot systems to achieve optimal application performance and stability
  • Collaborate effectively with development, testing and management teams in disparate locations.
  • Optimize the reliability and performance of our software solutions
  • Design and implement proactive monitoring to ensure health, performance and security of our production environment
  • Optimize application performance
  • Maintain configuration documentation
  • A self-starter with an inquiring and open mind, detail oriented, capable of stepping back and seeing a bigger picture
  • Demonstrated ability to collaborate and communicate clearly and effectively with people from both technical and non-technical areas, including client-facing communications and creation of formal documentation.
  • Strong understanding of how technology relates to business, market and industry and applies such knowledge to support the needs of the business.
  • Experience in developing and/or debugging application code and installing third party vendor software
  • Solid experience with Informatica, Oracle, Teradata, Unix, ETL, Power Exchange , PL/SQL uDeploy, Data analysis, Hadoop
  • Previous experience with specific enterprise computing tools: Service Now for Incident, Problem and Change, Connect: DIRECT (NDM) file transmission

Vacancy expired!

Subscribe Report job