6 + years of extensive IT experience with multinational clients with over 4+ years of Bigdata experience developing Bigdata/Hadoop applications.
Excellent understanding and knowledge of Hadoop, Spark architecture
Experience in AWS services such as EMR, EC2, S3, RedShift which provides fast and efficient processing of Big Data.
Experience in GCP service such as BigQuery, Dataproc,Dataflow,Dataprep etc.
Experience in importing and exporting terabytes of data using Sqoop from HDFS to Relational Database Systems and vice - versa.
Experienced in Hive partitioning and bucketing to optimize the performance.
Experienced with performing real-time analytics on NoSQL databases like MongoDB.
Used Flume to channel data from different sources to HDFS.
Utilized Kafka and Spark Streaming to stream data in near real-time into HDFS.
Worked on Airflow scheduler for job scheduling and also created custom operators in the dag for different tasks.
Involved in various projects related to Data Modeling, System/Data Analysis, Design and Development for both OLTP and Data warehousing environments using ER Studio.
Good understanding of the Data modeling (Dimensional & Relational) concepts like Star-Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables.