Hadoop Developer
Eden Prairie, MN
1 year
Required:
Technologies  | Mandatory(M)  | 
Java  |  Good to Have  | 
SQL  |  M  | 
Datastage  | M  | 
Hadoop  |  M  | 
Map Reduce  |  Good to Have  | 
Hive  |  M  | 
Pig  |  M  | 
Sqoop and MRUnit  | Good to Have  | 
Falcon  |  Good to Have  | 
UDF defined functions/ PIGUDF  | M   | 
HDFS file system  |  M  | 
Shell scripting  |  M  | 
Python  |  Good to Have  | 
Tivoli Work Scheduler or any scheduling Tool  | M  | 
 Eclipse  |  Good to Have  | 
WebSphere  |  Good to Have  | 
Oracle/TeraData or any other Database  | M  | 
Responsibilities: 
• Worked on live hadoop cluster running on MapR distribution. 
• Involved in creating Hive tables, loading with data and writing hive queries which will run internally in MapReduce way. 
• Used Sqoop to ingest data from various source systems into HDFS. 
• Written Hive queries for data analysis to meet the business requirements 
• Familiarized with job scheduling so that CPU time is well distributed amongst all the jobs. 
• Executed queries using Hive and developed Map-Reduce jobs to analyze data. 
• Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS. 
• Developed Hive queries for the analysts. 
• Worked on DataGuise and documented the same for internal users. 
• Worked on Elastic search and created various scripts. 
• Worked on Datameer and created the shell script to run the Datameer job from the command line. 
 
Environment: Eclipse, Oracle 10g, Hadoop, MapReduce, Hive, Linux, MapReduce, HDFS, Hive, MapR, SQL, Toad 9.6.
Posted by: "Eshwar Yerram" <eshwar@3ktechnologies.com>
| Reply via web post | • | Reply to sender | • | Reply to group | • | Start a New Topic | • | Messages in this topic (1) |