Vacancy expired!
Company Description
pureIntegration, a technology consulting firm with 17+ years of experience servicing fortune 100 clients is seeking a Framework Engineer that will be responsible for building frameworks for Big Data manipulation. This role will focus heavily on working with both framework and data management teams. The candidate for this role must be proficient in both the development, build, and support of big data frameworks ideally using Spark.Location: Philadelphia, PA. Work Arrangement: Contract W2 hourly role, initial contract period through 12/31/2022 with possible extension for additional year. Work Authorization: USCIT, Green Card. We Do Not Sponsor H1B Visas, or C2CResponsibilities:- Hands-on building of ETL pipelines using our internal framework written in Java
- Hands-on solutioning of real time REST APIs or other solutions for streaming data from GraphDB
- Modify existing application code or interfaces or build new application components from detailed requirements.
- Analysis of requirements, support of the design, development of the code, testing, debugging, deployment, and maintenance of those programs and interfaces. Documentation of the work is essential
- Participation in most aspects of programming and application development, including file design, update, storage, and retrieval
- Enhance processes to resolve operational problems and add new functions taking into consideration schedule, resource constraints, process complexity, dependencies, assumptions and application structure
- Ability to maintain the developed solution on an on-going basis is essential
- Ability to follow the existing development methodology and coding standards, and ensure compliance with the internal and external regulatory requirements
- Develop and implement databases, data collection systems, data analytics and other strategies that optimize statistical efficiency and quality
- Acquire data from primary or secondary data sources and maintain databases/data systems
- Work with team to plan Quarterly work, estimating story points and aligning project work with management to prioritize business and information needs
- Locate and define new process improvement opportunities
- Document design and data flow for existing and new applications being built.
- Co-ordinate with multiple different teams QA, Operations and other development team within the organization.
- Testing methods, including unit and integration testing (JUnit, JMock)
- Ability to integrate with large teams, demonstrating strong verbal and written communication skills
- Utilization of software configuration management tools
- Code deployment and code versioning tools
- Bachelor's degree in Engineering, Computer Science or a related field, or relevant work experience.
- At least 4+ years of experience implementing complex ETL pipelines (batch and streaming)
- At least 4+ years of experience with Java particularly within the data space
- At least 2+ years of experience w/ Spark (or similar Big Data)
- Must have experience in JAVA, J2EE, EJB, JavaScript, servlets, JDBC, struts, Hibernate
- Technical expertise regarding data models, database design development, data mining and segmentation techniques
- Good experience writing complex SQL and ETL processes
- Excellent coding and design skills in Java , Scala and Python are a plus
- Experience working with large data volumes, including processing, transforming and transporting large-scale data
- Experience in AWS technologies such as EC2, Redshift, Cloud formation, EMR, AWS S3, AWS Analytics required.
- Big data related technologies like HIVE, Presto, Apache Hadoop, Apache Spark, Kafka required.
- AWS certification is preferable: AWS Developer/Architect/Big Data
- Strong analytical skills with the ability to collect, organize, analyze, and disseminate significant amounts of information with attention to detail and accuracy
- Good Experience building reusable frameworks.
- Experience working in an Agile Team environment.
- Excellent communication skills both verbal and written
Vacancy expired!