Hadoop Lead
VenturesUnlimitedInc2
Posted: August 7, 2015
Interested in this position?
Create a free account to apply with AI-powered matching
Quick Summary
Design and implement map reduce jobs to support distributed processing using java, python, hive and pig; Ability to design and implement end to end solution.
Required Skills
Job Description
Ventures Unlimited is a service centric organization. Our success begins with our ability to apply our relationships, leverage technologies, and identify talent - often in combination - to an endless array of real-world customer needs
**** FULL TIME PERMANENT POSITION******
Position : Hadoop Lead with Python
Location: Beaverton, OR
Duration: Full Time
Job Description :
Responsibilities:
• Design and implement map reduce jobs to support distributed processing using java, python, hive and pig; Ability to design and implement end to end solution.
• Build libraries, user defined functions, and frameworks around Hadoop
• Research, evaluate and utilize new technologies/tools/frameworks around Hadoop eco system
• Develop user defined functions to provide custom hive and pig capabilities
• Define and build data acquisitions and consumption strategies
• Define & develop best practices
• Work with support teams in resolving operational & performance issues
• Work with architecture/engineering leads and other teams on capacity planning
• Work with Site-Operations team on configuration/upgrades of the cluster
• Excellent communication skills
Qualification:
• MS/BS degree in a computer science field or related discipline
• 8+ years’ experience in large-scale software development
• Experience with data warehousing, dimensional modeling and ETL development
• Strong Java programming, shell scripting, Python, and SQL
• Strong development skills around Hadoop, MapReduce, Hive, Pig, Impala
• Strong understanding of Hadoop internals
• Good understanding of AVRO and Json
• Experience with build tools such as Maven
• Experience with databases like Oracle;
• Experience with performance/scalability tuning, algorithms and computational complexity
• Experience (at least familiarity) with data warehousing, dimensional modeling and ETL development
• Ability to understand and ERDs and relational database schemas
• Proven ability to work cross functional teams to deliver appropriate resolution
• Experience with open source NOSQL technologies such as HBase and Cassandra
• Experience with messaging & complex event processing systems such as Kafka and Storm
• Machine learning framework (Nice to have)
• Statistical analysis with Python, R or similar
• MS/BS degree in a computer science field or related discipline
• 8+ years’ experience in large-scale software development
• Experience with data warehousing, dimensional modeling and ETL development
• Strong Java programming, shell scripting, Python, and SQL
• Strong development skills around Hadoop, MapReduce, Hive, Pig, Impala
Contact "Mohit Maurya" for more details at 856-842-1988 Ext:141