Professional Summary:
• 3.7 years of IT experience in Big Data and Hadoop Ecosystems.
• Experience in providing Big Data solutions using Hadoop and its components like HDFS, Sqoop, Flume, Hive, SparkSQL, Spark Streaming, Kafka, Cassandra, Scala, AWS, MapReduce, Yarn, Apache Drill, Hbase, Oozie
• Experience in processing real time streaming data using AWS ecosystem like Kinesis, IoT Core, EMR, Lambda, S3, Glue, RDS, EC2, IAM, DynamoDB, Elastic Cache.
• Written Hive queries for creating managed/external tables, data preprocessing, data transformation and analysis.
• Implemented partitioning, bucketing in Hive to optimize performance.
• Importing and exporting data into HDFS from RDBMS and vice versa using Sqoop.
• Worked with RDDs, dataframes, datasets, SparkSQL using Scala.
• Queried Hive tables using SparkSQL.
• Good understanding of Hadoop architecture, MapReduce programming model, HDFS, various file formats.
• Optimized Spark and Hive using various optimization techniques.
Would like to hear from you for this opportunity.
I am a quick learner so will learn some skills on the fly.