TopDevz · Nov 3rd 2020
We are looking for an experienced, senior, Business Analyst and Big Data Architect, who is excited to work on one of our many client projects - both greenfield (new) projects as well as legacy (support) projects in that technology stack. This is a remote position, 40 hours a week.
Skills & Requirements
Expert level knowledge with 8-10 years of experience in Cloudera Hadoop components such as HDFS, HBase, Impala, Hue, Spark, Hive, Kafka, YARN and ZooKeeper
Expertise in architecture, Build and Troubleshooting Hadoop experience on Cloudera Distribution
Hands on experience with ON prem and cloud deployments
Validated experience in scripting, automation, deployment, set up and installation, trouble shooting and fixing issues across platforms
Architecture, Design and Development of Big Data lake
Take end to end responsibility of the Hadoop life Cycle in the organization.
Detect, analyze and remediate performance problems.
3-5 years of experience in at least one of the following: Python, Java or Shell Scripting and eager to pick up new programming languages on the go
The ability to function within a multidisciplinary, global team. Be a self-starter with a strong curiosity for extracting knowledge from data and the ability to elicit technical requirements from a non-technical audience
Data Concepts (ETL, near-/real-time streaming, data structures, metadata and workflow management)
You have deep understanding of Dev ops and Agile software development methodologies
You have strong communication skills and the ability to present deep technical findings to a business audience
Hands-on experience working in an agile environment
AWS/Azure/Google Cloud experience is a plus.
Experience in architecture and troubleshooting skills on tech stacks HDFS,HIVE, Mapreduce, YARN,IMPALA,SPARK,HBASE, KMS, HUE, OOZIE, SENTRY
Architecting Large Scale Big Data/Analytics solutions
Architecture/Design Patterns for Big Data and Cloud systems
Build analytics tools that utilize the data pipeline to provide actionable insights into key business performance metrics.
Assemble large, complex data sets that meet functional / non-functional business requirements.
Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc
Responsible for defining the Build and Manage part of Hadoop solutions