• Extensive experience in Big Data ecosystem and its various components such as SPARK, SCALA, MapReduce, HDFS, HIVE, HBase, Sqoop, Zookeeper, Kafka and Flume,Java.
• Good understanding/knowledge of Hadoop Architecture and its components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, Secondary Name node, and MapReduce concepts.
• Experience in handling different file formats like Text files, Sequence files, Avro data files using different SerDe's in Hive.
• Hands on to connect & read data from Hive & MySql using PrestoDB with PHP client.
• Experienced in performing analytics on structured and unstructured data using Hive queries.
• Experience in process improvement, Normalization/de-Normalization, data extraction, data cleansing, data manipulation on HIVE.
•Good exposure on usage of NoSQL database.