Вы находитесь на странице: 1из 3

SHUBHAM ROKADE

shubham.rokade12@gmail.com
+91 9404762058 | 9503412453

OBJECTIVE
To continuously enhance my knowledge, skill and experience by getting involved in challenging
work environment
nvironment and utilize them for personal and organizational goal to the best of my ability.

CARRIER SUMMARY
 Hands on experience with Big Data core components and Eco System including Data Ingestion
and Data Processing
cessing (Spark Scala, Hive, Sqoop,
Sqoop HDFS and MapReduce )
 Experience Installing ,Configuring and Testing Hadoop Ecosystem Components
Component
 In-depth
depth understanding of Spark Architecture including
inc Spark-Core,
Core, Spark
Spark-Sql, Dataframes,
Basics of Spark-Streamimg.
 Strong Knowledge of Hadoop and Hive and Hive's analytical functions.
 Capable of processing large sets of structured, semi-structured
semi structured and unstructured data.
 Capturing data from existing databases that provide SQL interfaces using Sqoop.
 Implemented Proofs of Concept on Hadoop stack and different big data analytic tools, Migration
from different databases (i.e Teradata, Oracle,MYSQL ) to Hadoop.
 Loaded the dataset into Hive for ETL Operation.
Opera
 Very Good understanding of Partitioning, Bucketing concepts in Hive and designed both Internal
and External tables in Hive to optimize performance.
 Solved performance issues in Hive scripts with understanding of Joins, Group and Aggregation.
 Hands on n experience in IDE tools like Eclipse,IntelliJ,QUBOLE
 Good exposures in Requirement Analysis, problem solving skills, good communication and
interpersonal skills.
 Received appreciation and recognition from client several times.

TECHNICAL SKILL
 Big Data Ecosystem : Hadoop, MapReduce, Hive,Impala, HBase, Sqoop, Spark, Scala
 Scripting Language: Python and Shell Scripting
 Programing Language:C,C++,
uage:C,C++, Python, Java, SQL,
 Operating System:Linux-Ubuntu
Ubuntu 16.04,CentOS,Windows7/8/10.
16.04
 Databases: Oracle MySql
 Cloud Platform: AWS S3

PROFESSIONAL EXPERIENCE
EC-Mobility, Aurangabad, India Jun 2017--Dec
2017 2019

Data Engineer
WORK EXPERIENCE

Project : TUI Future Market. Jul 2019--Dec 2019

Technologies: Spark Core, Spark-SQL, Scala, JIRA, QUBOLE, AWS S3.

The basic agenda of this project is for price comparison analysis on the basis of various tables. We have
used AWS as a cloud platform and with the help of QUBOLE we did the processing. also datalake which was
on to the cloud itself,using various sources we have processd the data and convert them into a meaningful
insight.

Roles and Responsibilities:


• Fetch data in Qubole Analysis Platform
• Implemented RDD to DataFrame Transformation,when the data in txt format.
• Implemented some Performance Optimization Techniques.
• Stored data back to respective AWS S3 location

Project : Data Ingestion and Data Lake. Jan 2018--Jul 2019

Technologies: CDH, HDFS, MapReduce, Hive, Sqoop, Shell scripting, JIRA.

The purpose of the project is to do the ingestion of the data that is present in Mainframe system to
Big Data Lake. It’s Mainframe to hadoop migration project. History creation process of data
warehouse application and perform aggregation operations on Hadoop Big Data platform.

Roles and Responsibilities:


• Involved in Collecting Business Requirements from Business Users, Translate into
Technical Design (Data Pipelines and ETL workflows).
• Import Data into the Hive from various Relational Databases (Oracle, Sybase, SQL
Server) using Sqoop.
• Write Hive DDL to Create Hive Table for Optimize Query Performance.
• Ingest Flat Files like Delimited, Fixed Length, etc. into Hive Warehouse.
• Define all the possible Test Cases along with the Test Data.
• Involved in import data from various RDBMS into HDFS using Sqoop which
includes Incremental Load to populate Hive External Table and vice-versa.
• Designed both Managed and External Hive Tables and Defined static and dynamic
partitions as per requirement for optimized performance on production datasets.
• Worked with various File Formats like Text File, SequenceFile, ORC Files, Avro Files
and various Compression Formats like Snappy, bzip2.
• Written Hive Queries for Data Analysis to meet the business requirements.
• Built reusable Hive UDF libraries for business requirements which enabled users to
use these UDF's in Hive Querying.
• Implement Data Validation and Quality Checks.
EDUCATION

2013-2017 SSBT’s College of Engineering & Technology, Jalgaon


Bachelor of Engineering CGPA : 7.71
(Computer Engineering)

June-2013 Late G.D.Mali Junior College, Deopur-Dhule


HSC (Science) Percentage : 60.67 %

May-2011 S.T.T.K Mahajan Highschool Deopur-Dhule


SSC Percentage : 79.20 %

PERSONAL DETAILS
Name : Shubham Madhukar Rokade

Father’s Name : Madhukar Ramdas Rokade

Mother’s Name : Sunita Madhukar Rokade

Date of Birth : 12 December 1994.

Address : Padmashri Residency-II, Flat No.8, Sant Gadage Baba Colony,

Datta Mandir, Deopur-Dhule (MS).Pin-424005

Gender : Male.

Blood Group : O+

Hobbies : Caricature, Drawing, Badminton

Strength : Dedicated towards Work.

DECLARATION
I hereby declare that the above-mentioned information is true and to the best of my knowledge.
Date :

Place :
Shubham Madhukar Rokade

Вам также может понравиться