Master’s degree in Computer Science, Engineering (any), Technology, Management Information Systems or related and 2 yrs of exp to Responsible for capacity planning and estimating the requirements for lowering or increasing the capacity of the Hadoop cluster. Monitoring and Maintenance of production cluster. Import logs from Web servers with Flume to ingest the data into HDFS. Involve in extracting the data from various sources into Hadoop HDFS for processing. Troubleshooting backup and restore problems and constant monitoring of system performance. Develop Map-Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.