Vacancy expired!
Title : Big Data Engineer with strong Scala experience is must preferred local onlyLocation : San Francisco, CADuration:12+ monthsJob description:10+ years of demonstrable experience designing technological solutions to complex data problems, developing & testing modular, reusable, efficient and scalable code to implement those solutions. Ideally, this would include work on the following technologies:Expert-level proficiency in Scala/PysparkStrong understanding and experience in distributed computing frameworks, particularly Apache Hadoop (YARN, MR, HDFS) and associated technologies one or more of Hive, Sqoop, Avro, Flume, Oozie, Zookeeper, Impala, etc.Hands-on experience with Apache Spark and its components (Streaming, Core, DF, RDD, SQL, MLLib) is a strong advantage.Operating knowledge of cloud computing platforms (AWS/AzureP)Experience working within a Linux computing environment, and use of command line tools including knowledge of Shell/Python scripting for automating common tasksAbility to work in a team in an agile setting, familiarity with JIRA and clear understanding of how Git works and code automation to build CI/CD deployment pipelineIn addition, the ideal candidate would have great problem-solving skills, and the ability & confidence to hack their way out of tight corners.ExperienceMust Have (hands-on) experience:Scala or Python expertiseLinux environment and shell scriptingDistributed computing frameworks (Hadoop or Spark)Cloud computing platforms (AWS/AzureP)Desirable (would be a plus):Statistical or machine learning DSL like RDistributed and low latency (streaming) application architectureRow store distributed DBMSs such as CassandraFamiliarity with API design & real-time processing tool like Kafka.
Vacancy expired!