1 Click Easy Apply


Solution Architect - Hadoop Development


Minneapolis, Minnesota


Solution Architect - Hadoop Development Job Opening in Minneapolis, Minnesota - Solution Architect (SA) ? Hadoop Architecture & Development
Big Data is all we do.
At phData, our Consulting Services Team is responsible for developing world-class big data solutions. As a Solution Architect, you will partner with key client stakeholders on a variety of Hadoop projects to transform complex data problems into innovative results.
For more information, check out our website at
At phData, Solution Architects are transformative leaders in Hadoop Technologies
Solution Architects are responsible for determining how the technical project will be executed including selecting the best tools with overall responsibilities for performance, data integration, ecosystem integration and security as well as general project management oversight. Solution Architects work across a broad range of technologies ? from infrastructure to applications ? to ensure the ideal solution is implemented and optimized. Work with key business stakeholders (BA, Product Owner, PM and internal IT teams) as well as coach and mentor Data Engineers. Architects will work with 2+ simultaneous clients or multiple projects within a client.
Responsibilities include:
Integrate data from a variety of data sources (data warehouse, data marts) utilizing on-prem or cloud-based data structures (AWS); determine new and existing data sources
Design and implement streaming, data lake, and analytics big data solutions
Create and direct testing strategies including unit, integration, and full end-to-end tests of data pipelines
Choose the right storage solution for a project. Comparing Kudu, HBase, HDFS, and relational databases based on their strengths
Utilize ETL processes to build data repositories; integrate data into Hadoop data lake using Sqoop (batch ingest), Kafka (streaming), Spark, Hive or Impala (transformation)
Ability to learn new technologies in a quickly changing field
Partner with our Managed Services team to design and install on prem or cloud based infrastructure including networking, virtual machines, containers, and software
Determine and select best tools to ensure optimized data performance; perform Data Analysis utilizing Spark, Hive, and Impala
Strong working knowledge of SQL and the ability to write, debug, and optimize distributed SQL queries
Mentor and coach developers and data engineers. Provide guidance with project creation, application structure, automation, code style, testing, and code reviews
Qualifications
Previous experience as a Software Engineer, Data Engineer or Data Analytics
Expertise in core Hadoop technologies including HDFS, Hive and YARN.
Deep experience in one or more ecosystem products/languages such as HBase, Spark, Impala, Solr, Kudu, etc
Expert programming experience in Java, Scala, or other statically typed programming language
Excellent communication skills including proven experience working with key stakeholders and customers
Ability to translate ?big picture? business requirements and use cases into a Hadoop solution, including ingestion of many data sources, ETL processing, data access and consumption, as well as custom analytics
Experience scoping activities on large scale, complex technology infrastructure projects
Customer relationship management including project escalations, and participating in executive steering meetings
Our Team: Accelerate your growth and success as you partner with some of the brightest minds in the industry. As the world?s largest pure-play Big Data services firm, our team includes Apache committers, Spark experts and the most knowledgeable Scala development team in the industry. phData has earned the trust of customers by demonstrating our mastery of Hadoop services and our commitment to excellence.
Our Approach: Creative latitude; work across a broad range of big data technologies ? from infrastructure to applications ? we ensure the ideal solution is implemented and optimized.
Keywords:
MapReduce, HBase, Hive, Apache Spark, Java, Apache Kafka, Big Data, Spark, Solution Architecture, Cloudera, Apache Pig, Hadoop, NoSQL, Cloudera Impala, Scala, Python, Data Engineering, Big Data Analytics, Large Scale Data Analysis, ETL, Linux, Kudu (not available in LI)
Job Type: Full-time
Required education:
Bachelor's
Required experience:
Hadoop Technologies: 3 years
Hadoop Development: 3 years
Java: 3 years
Software Engineering: 5 years

1 Click Easy Apply

TalentEinstein.com - Superhuman AI Recruiting Assistant | Terms & Conditions

All rights reserved
Swanco LLC