Vacancy expired!
- Involved in the analysis of unstructured and semi-structured data, including latent semantic indexing (LSI), entity identification and tagging, complex event processing (CEP), and the application of analysis algorithms on distributed, clustered, and cloud-based high-performance infrastructures.
- Exercises creativity in applying non-traditional approaches to large-scale analysis of unstructured data in support of high-value use cases visualized through multi-dimensional interfaces.
- Handle processing and index requests against high-volume collections of data and high-velocity data streams.
- Has the ability to make discoveries in the world of big data.
- Requires strong technical and computational skills - engineering, physics, mathematics, coupled with the ability to code design, develop, and deploy sophisticated applications using advanced unstructured and semi-structured data analysis techniques and utilizing high-performance computing environments.
- Has the ability to utilize advance tools and computational skills to interpret, connect, predict and make discoveries in complex data and deliver recommendations for business and analytic decisions.
- Experience with software development, either an open-source enterprise software development stack (Java/Linux/Ruby/Python) or a Windows development stack (.NET, C#, C).
- Participates as a member of a Data Engineering Team supporting one or more Agile application teams. Works collaboratively with the Application and Infrastructure teams, including architects, developers and testers, to ensure project success. Participates in database development and maintenance activities including:
- develop, modify, test, release/deploy and maintain simple database objects
- maintain database security and audit trails
- assist with execution or testing of database recovery
- monitor databases
- assist with DBMS SW patching or upgrades
- Ability to develop descriptive, predictive, or prescriptive analytics to make sense of structured and unstructured data.
- Knowledge of cloud/virtual content architectures, data standards & workflows, and best practices.
- Experience in system activity and data modeling (ERWin) or others, information flow (Visio)/ transactional process analysis, business methods, and performance measurements techniques (Data Analytics & presentation tools (such as Tableau).
- Strong oral, writer, and presentation skills. Involved in the development of complex SQL queries and scripts for use in Ad Hoc Reports, analysis of data quality, and data manipulation within an Oracle database.
- Strong data analytical skills and experience with the ability to communicate is necessary.
- Familiarity with working on fast-paced Agile projects.
- Experience with data transport and transformation APIs and technologies such as JSON, XML, XSLT, JDBC, SOAP and REST. Experience with Cloud-based data analysis tools including Hadoop and Mahout, Acumulo, Hive, Impala, Pig, and similar. Experience with visual analytic tools like Microsoft Pivot, Palantir, or Visual Analytics. Experience with open source textual processing such as Lucene, Sphinx, Nutch or Solr. Experience with entity extraction and conceptual search technologies such as LSI, LDA, etc. Experience with machine learning, algorithm analysis, and data clustering
- Requires 8 to 10 years with BS/BA or 6 to 8 years with MS/MA or 3 to 5 years with PhD.
- Relevant Data Scientist / Architect / Engineering Professional credentials, academic certification, and /or training
- Prefer or more of these certifications: INCOSE, ITIL, CISSP, Agile
- Familiarity with working on SAFe agile projects
Vacancy expired!