SlideShare une entreprise Scribd logo
1  sur  3
Sanath Pabba
Mobile: +1-4168939098 Email: sanath.sanny@gmail.com Skype id : live:sanath.sanny_1
Synopsis
 Engineering Professional with 5+ years of experience in Development, Maintenance and Production
support in Bigdata technologies.
 WorkingdirectlywiththeClientsregardingcritical requirementsandindependently workedonManyapps
and provided many ideas to improve the system efficiency.
 Written 12 automated scripts that optimized the duties and saved 70-80 hours of efforts per month.
 Out of 12 automations,4 automationsgiven8-10% of data improvementswhichhelpsdatapreparation
teams effort.
 Expertise with the tools in Hadoop Ecosystem including Pig, Hive, HDFS, MapReduce, Sqoop, Storm,
Spark, Kafka, Yarn, Oozie, and Zookeeper.
 Excellent knowledge on Hadoop Ecosystemssuch as HDFS, Job Tracker, Task Tracker, Name Node, Data
Node and Map Reduce programming paradigm.
 Experience indesigninganddevelopingapplicationsin SparkusingScala to compare the performance of
Spark with Hive and SQL/Oracle.
 Good exposure with Agile software development process.
 Experience in manipulating/analysing large datasets and finding patterns and insights within structured
and unstructured data.
 Strong experience on Hadoop distributions like Cloudera, MapR and HortonWorks.
 Good understandingof NoSQLdatabasesandhandsonworkexperience inwritingapplicationsonNoSQL
databases like HBase, and MongoDB.
 Experienced in writing complex MapReduce programs that work with different file formats like Text,
Sequence, Xml, parquet, and Avro.
 Experience in Oozie and workflow scheduler to manage Hadoop jobs by Direct Acyclic Graph (DAG) of
actions with control flows.
 Experience in migrating the data using Sqoop from HDFS to Relational DBMS and vice-versa.
 Extensive Experience onimportandexportdatausingstreamprocessingplatformslikeFlume andKafka.
 Very good experience in complete project life cycle (design, development,testing and implementation)
of Client Server and Web applications.
 Strong knowledge on Data Warehousing ETL concepts using Informatica Power Center, OLAP, OLTP.
 Having expertise knowledge on banking, insurance, and manufacturing domains.
 Excellentproblem-solvingskillswithastrong technical backgroundand resultorientedteamplayerwith
excellent communication and interpersonal skills.
 Worked on creating the RDDs, DFs for the required input data and performed the data transformations
using Spark-core.
 Hands-on experience in Scala programming and Spark components like Spark-core and Spark-SQL.
Technical Skill Set
 BigData Technologies:HDFS,MapReduce,Hive,Pig,Sqoop,Flume,Spark,Kafka,ZookeeperandOozie
 Languages: C, Java,Scala,Python,SQL,PL/SQL, PigLatin,HiveQL, JavaScript,Shell Scripting
 DevelopmentMethodologies:Agile/Scrum, Waterfall VersionControlToolsGit,SVN,Bitbucket
 RDBMS: Oracle,SQL Server
 BuildTools: Jenkins,Maven,ANT
 BusinessIntelligence Tools:Tableau,Splunk,QlikView,Alteryx
 Tools: IntelliJIDE
 CloudEnvironment:AWS
 Scripting: Unix shell scripting,Pythonscripting
 Scheduling:Maestro
Career Highlights
ProficiencyForte
 Extensivelyworkedondataextraction,Transformationandloadingdatafromvarioussourceslike
DB2, Oracle and Flat files.
 StrongskillsinData RequirementAnalysisandDataMappingfor ETL processes.
 Well versedindevelopingthe SQLqueries,unions,andmultiple table joins.
 Well versedwithUNIXCommands&able towrite shellscriptsanddevelopedfew scriptsto reduce
the manual interventionaspartof JobMonitoringAutomationProcess.
Work experience
 Sep2019 till currentdate in WalmartCanada as CustomerExperienceSpecialist(Part-time).
 Jan 2018 to Mar 2019 in Infosys Limitedas SeniorSystemEngineer.
 Apr 2015 to Jan2018 in NTT Data Global DeliveryServices asApplicationsoftware dev.Consultant.
Project Details
Company : InfosysLimited.
Project : Enterprise businesssolution(Metlife Inc.)
Environment : Hadoop, Spark,Spark SQL, Scala,SQL Server, shell scripting.
Scope: EBS is a project where Informatica pulls data from SFDC and sends to Big Data at RDZ. Big Data
kicksitsprocesswhenthe triggerfile,control fileanddatafilesare received.Allthe filescheckvalidations.
Afterall the transformationsare done,the data is storedin hive,pointingtoHDFS locations.The data is
synced to bigsql and down streaming process is done by QlikView team.
Roles& Responsibilities:
 Writing Sqoop Jobs that loads data from DBMS to Hadoop environments.
 Preparedcode thatinvokessparkscriptsinScalacode thatinvolvesinDataloads,pre-validations,data
preparation and post validations.
 Prepared automation scripts using shell scripting that fetches the data utilizations across the cluster
and notifies admins for every hour that helps admin team to avoid regular monitoring checks.
Company : InfosysLimited.
Project : BluePrism(MetlifeInc.)
Environment : Spark,Spark SQL, Scala,Sqoop,SQL Server,shell scripting.
Scope: BluePrism is a Source application with SQL Server as its Database. Big Data will extract the data
fromBluePrism Environmentsmergetwosourcesintooneandloadthe dataintoHive Database.BigData
also archives the data into corresponding history tables either monthly or ad-hoc basis basedon trigger
file receivedfromBluePrism.ThisisaweeklyextractfromSQLserverusingSqoopandthenloadthe data
through Scala. Jobs have been scheduled in Maestro.
Roles& Responsibilities:
 Prepared data loading scripts using shell scripting which invokes Sqoop jobs.
 Implemented data merging functionality which pulls data from various environments.
 Developed scripts that backup data using AVRO technique.
Company : Infosys Limited.
Project : Gross ProcessingMarginreports(Massmutual)
Environment : Spark,Spark SQL, Scala,Sqoop,SQL Server,shell scripting.
Scope:In GPMReports,we receive the input.csvfromthe Business. Basedonthe clientrequestwe need
to generate 6 reports.We will receive the triggerfile anddata file for each report.Using shell script,we
will perform validationon the trigger files, input file and paths that representation in Linux and HDFS, if
everyvalidationsuccessfulandinvoke the hive scripttogenerate the outputfile andplacedinLinux.We
will append the data in hive tables based on the output file. Migrating the Pig Scripts into spark scripts
using Scala and report generation will be taken place by the spark and stores in Linux directory.
Roles& Responsibilities:
 Based on the business need, used to prepare the data using hive QL and Spark RDD’s.
 By using Hadoop, we load the data, prepare the data, implement filters to remove unwanted and
uncertain fields and merging all 6 reports from various teams.
 Implemented8pre-validationrulesand7 postvalidationruleswhichinvolvesindata count,required
fields and needful changes and in post validations we move the data to HDFS archive path.
Company : Ntt Data Global deliveryservices.
Project : Compliance Apps(National Lifegroup)
Environment : Spark,Spark SQL, Scala,SQL Server,shell scripting,pig.
Scope:Compliance appisagroupof nine adminsystems(AnnuityHost,ERL,MRPS,CDI,Smartapp,WMA,
VRPS,PMACS,SBR).The processistoload the datafilesbasedonthe triggerfileswe receivedforthe nine
admin systems in to HDFS. There are three types of Load that takes place. They are:
1. Base load
2. Full load
3. Deltaload
Tobuildthe workflowtoloadthe datafilesintoHDFSlocationsandintohive tables.We needtocreate
hive tableswithoptimizedcompressedformatandload the data into the tables.To write hive script for
full load and write the shell script to create a workflow. We use Pig/spark for the delta loads and shell
script to invoke the hive for the full load/Historyprocessing.Then,schedule the jobsinMaestro for the
daily run. Initially for delta load, we were using Pig scripts.
Company : NTT Data Global DeliveryServices.
Project : ManufacturingCostWalk Analysis(Honeywell)
Environment : Sqoop,Shell scripting,Hive.
Scope: The Manufacturingcost walkapplicationusedtostore the informationaboutthe productswhich
are beenmanufacturedbyHoneywell.Theyusedtostore the datainSharePointlistsonweeklybasis.But
still it is very difficult for them to handle the data using share point because of long time processing. So
we proposeda solutionforthemwithHive andSqoop.But theirsource of file generationisfromcsvand
xlsfiles.So,we have startedimporting data into hive and processing data based on their requirement.
Academia
 Completed Post graduation from Loyalist college in Project Management (Sep 19- May 20)
 Completed diploma from Indian institute of Information Technology (Hyderabad) in Artificial
Intelligence and data visualization (April 2019 – Aug 2019)
 Completed Graduation Under JNTU-HYD in Electronics and Communications (Aug 11 – May 14)

Contenu connexe

Tendances

a9TD6cbzTZotpJihekdc+w==.docx
a9TD6cbzTZotpJihekdc+w==.docxa9TD6cbzTZotpJihekdc+w==.docx
a9TD6cbzTZotpJihekdc+w==.docxVasimMemon4
 
Introduction To Big Data with Hadoop and Spark - For Batch and Real Time Proc...
Introduction To Big Data with Hadoop and Spark - For Batch and Real Time Proc...Introduction To Big Data with Hadoop and Spark - For Batch and Real Time Proc...
Introduction To Big Data with Hadoop and Spark - For Batch and Real Time Proc...Agile Testing Alliance
 
R and Big Data using Revolution R Enterprise with Hadoop
R and Big Data using Revolution R Enterprise with HadoopR and Big Data using Revolution R Enterprise with Hadoop
R and Big Data using Revolution R Enterprise with HadoopRevolution Analytics
 
Big Data Simplified - Is all about Ab'strakSHeN
Big Data Simplified - Is all about Ab'strakSHeNBig Data Simplified - Is all about Ab'strakSHeN
Big Data Simplified - Is all about Ab'strakSHeNDataWorks Summit
 
Big dataarchitecturesandecosystem+nosql
Big dataarchitecturesandecosystem+nosqlBig dataarchitecturesandecosystem+nosql
Big dataarchitecturesandecosystem+nosqlKhanderao Kand
 
Predictive Analytics with Hadoop
Predictive Analytics with HadoopPredictive Analytics with Hadoop
Predictive Analytics with HadoopDataWorks Summit
 
Big data processing with apache spark
Big data processing with apache sparkBig data processing with apache spark
Big data processing with apache sparksarith divakar
 
Srikanth hadoop 3.6yrs_hyd
Srikanth hadoop 3.6yrs_hydSrikanth hadoop 3.6yrs_hyd
Srikanth hadoop 3.6yrs_hydsrikanth K
 
Real World Use Cases: Hadoop and NoSQL in Production
Real World Use Cases: Hadoop and NoSQL in ProductionReal World Use Cases: Hadoop and NoSQL in Production
Real World Use Cases: Hadoop and NoSQL in ProductionCodemotion
 

Tendances (20)

hadoop exp
hadoop exphadoop exp
hadoop exp
 
a9TD6cbzTZotpJihekdc+w==.docx
a9TD6cbzTZotpJihekdc+w==.docxa9TD6cbzTZotpJihekdc+w==.docx
a9TD6cbzTZotpJihekdc+w==.docx
 
Pallavi_Resume
Pallavi_ResumePallavi_Resume
Pallavi_Resume
 
Prashanth Kumar_Hadoop_NEW
Prashanth Kumar_Hadoop_NEWPrashanth Kumar_Hadoop_NEW
Prashanth Kumar_Hadoop_NEW
 
PRAFUL_HADOOP
PRAFUL_HADOOPPRAFUL_HADOOP
PRAFUL_HADOOP
 
Resume
ResumeResume
Resume
 
PRAFUL_HADOOP
PRAFUL_HADOOPPRAFUL_HADOOP
PRAFUL_HADOOP
 
Nagesh Hadoop Profile
Nagesh Hadoop ProfileNagesh Hadoop Profile
Nagesh Hadoop Profile
 
Introduction To Big Data with Hadoop and Spark - For Batch and Real Time Proc...
Introduction To Big Data with Hadoop and Spark - For Batch and Real Time Proc...Introduction To Big Data with Hadoop and Spark - For Batch and Real Time Proc...
Introduction To Big Data with Hadoop and Spark - For Batch and Real Time Proc...
 
Resume - Narasimha Rao B V (TCS)
Resume - Narasimha  Rao B V (TCS)Resume - Narasimha  Rao B V (TCS)
Resume - Narasimha Rao B V (TCS)
 
R and Big Data using Revolution R Enterprise with Hadoop
R and Big Data using Revolution R Enterprise with HadoopR and Big Data using Revolution R Enterprise with Hadoop
R and Big Data using Revolution R Enterprise with Hadoop
 
spark_v1_2
spark_v1_2spark_v1_2
spark_v1_2
 
Big Data Simplified - Is all about Ab'strakSHeN
Big Data Simplified - Is all about Ab'strakSHeNBig Data Simplified - Is all about Ab'strakSHeN
Big Data Simplified - Is all about Ab'strakSHeN
 
Big dataarchitecturesandecosystem+nosql
Big dataarchitecturesandecosystem+nosqlBig dataarchitecturesandecosystem+nosql
Big dataarchitecturesandecosystem+nosql
 
Predictive Analytics with Hadoop
Predictive Analytics with HadoopPredictive Analytics with Hadoop
Predictive Analytics with Hadoop
 
Big data processing with apache spark
Big data processing with apache sparkBig data processing with apache spark
Big data processing with apache spark
 
Mukul-Resume
Mukul-ResumeMukul-Resume
Mukul-Resume
 
Srikanth hadoop 3.6yrs_hyd
Srikanth hadoop 3.6yrs_hydSrikanth hadoop 3.6yrs_hyd
Srikanth hadoop 3.6yrs_hyd
 
Real World Use Cases: Hadoop and NoSQL in Production
Real World Use Cases: Hadoop and NoSQL in ProductionReal World Use Cases: Hadoop and NoSQL in Production
Real World Use Cases: Hadoop and NoSQL in Production
 
Resume_VipinKP
Resume_VipinKPResume_VipinKP
Resume_VipinKP
 

Similaire à Sanath pabba hadoop resume 1.0

Anil_BigData Resume
Anil_BigData ResumeAnil_BigData Resume
Anil_BigData ResumeAnil Sokhal
 
Rajeev kumar apache_spark & scala developer
Rajeev kumar apache_spark & scala developerRajeev kumar apache_spark & scala developer
Rajeev kumar apache_spark & scala developerRajeev Kumar
 
HariKrishna4+_cv
HariKrishna4+_cvHariKrishna4+_cv
HariKrishna4+_cvrevuri
 
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_Spark
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_SparkSunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_Spark
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_SparkMopuru Babu
 
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_Spark
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_SparkSunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_Spark
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_SparkMopuru Babu
 
Vishnu_HadoopDeveloper
Vishnu_HadoopDeveloperVishnu_HadoopDeveloper
Vishnu_HadoopDevelopervishnu ch
 
Manikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam M
 
Enterprise Data Workflows with Cascading and Windows Azure HDInsight
Enterprise Data Workflows with Cascading and Windows Azure HDInsightEnterprise Data Workflows with Cascading and Windows Azure HDInsight
Enterprise Data Workflows with Cascading and Windows Azure HDInsightPaco Nathan
 

Similaire à Sanath pabba hadoop resume 1.0 (20)

Monika_Raghuvanshi
Monika_RaghuvanshiMonika_Raghuvanshi
Monika_Raghuvanshi
 
Anil_BigData Resume
Anil_BigData ResumeAnil_BigData Resume
Anil_BigData Resume
 
Rajeev kumar apache_spark & scala developer
Rajeev kumar apache_spark & scala developerRajeev kumar apache_spark & scala developer
Rajeev kumar apache_spark & scala developer
 
resumePdf
resumePdfresumePdf
resumePdf
 
Arindam Sengupta _ Resume
Arindam Sengupta _ ResumeArindam Sengupta _ Resume
Arindam Sengupta _ Resume
 
sudipto_resume
sudipto_resumesudipto_resume
sudipto_resume
 
Sidharth_CV
Sidharth_CVSidharth_CV
Sidharth_CV
 
Sandish3Certs
Sandish3CertsSandish3Certs
Sandish3Certs
 
HariKrishna4+_cv
HariKrishna4+_cvHariKrishna4+_cv
HariKrishna4+_cv
 
Started with-apache-spark
Started with-apache-sparkStarted with-apache-spark
Started with-apache-spark
 
Vijay
VijayVijay
Vijay
 
RESUME_N
RESUME_NRESUME_N
RESUME_N
 
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_Spark
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_SparkSunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_Spark
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_Spark
 
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_Spark
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_SparkSunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_Spark
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_Spark
 
Vishnu_HadoopDeveloper
Vishnu_HadoopDeveloperVishnu_HadoopDeveloper
Vishnu_HadoopDeveloper
 
Manikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam_Hadoop_5+Years
Manikyam_Hadoop_5+Years
 
Resume (1)
Resume (1)Resume (1)
Resume (1)
 
Prasanna Resume
Prasanna ResumePrasanna Resume
Prasanna Resume
 
Enterprise Data Workflows with Cascading and Windows Azure HDInsight
Enterprise Data Workflows with Cascading and Windows Azure HDInsightEnterprise Data Workflows with Cascading and Windows Azure HDInsight
Enterprise Data Workflows with Cascading and Windows Azure HDInsight
 
Big data with java
Big data with javaBig data with java
Big data with java
 

Dernier

Accredited-Transport-Cooperatives-Jan-2021-Web.pdf
Accredited-Transport-Cooperatives-Jan-2021-Web.pdfAccredited-Transport-Cooperatives-Jan-2021-Web.pdf
Accredited-Transport-Cooperatives-Jan-2021-Web.pdfadriantubila
 
April 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's AnalysisApril 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's Analysismanisha194592
 
Chintamani Call Girls: 🍓 7737669865 🍓 High Profile Model Escorts | Bangalore ...
Chintamani Call Girls: 🍓 7737669865 🍓 High Profile Model Escorts | Bangalore ...Chintamani Call Girls: 🍓 7737669865 🍓 High Profile Model Escorts | Bangalore ...
Chintamani Call Girls: 🍓 7737669865 🍓 High Profile Model Escorts | Bangalore ...amitlee9823
 
VidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptxVidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptxolyaivanovalion
 
Log Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptxLog Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptxJohnnyPlasten
 
Invezz.com - Grow your wealth with trading signals
Invezz.com - Grow your wealth with trading signalsInvezz.com - Grow your wealth with trading signals
Invezz.com - Grow your wealth with trading signalsInvezz1
 
VIP Model Call Girls Hinjewadi ( Pune ) Call ON 8005736733 Starting From 5K t...
VIP Model Call Girls Hinjewadi ( Pune ) Call ON 8005736733 Starting From 5K t...VIP Model Call Girls Hinjewadi ( Pune ) Call ON 8005736733 Starting From 5K t...
VIP Model Call Girls Hinjewadi ( Pune ) Call ON 8005736733 Starting From 5K t...SUHANI PANDEY
 
FESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdfFESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdfMarinCaroMartnezBerg
 
100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptxAnupama Kate
 
Data-Analysis for Chicago Crime Data 2023
Data-Analysis for Chicago Crime Data  2023Data-Analysis for Chicago Crime Data  2023
Data-Analysis for Chicago Crime Data 2023ymrp368
 
CebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptxCebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptxolyaivanovalion
 
Week-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionWeek-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionfulawalesam
 
BigBuy dropshipping via API with DroFx.pptx
BigBuy dropshipping via API with DroFx.pptxBigBuy dropshipping via API with DroFx.pptx
BigBuy dropshipping via API with DroFx.pptxolyaivanovalion
 
Zuja dropshipping via API with DroFx.pptx
Zuja dropshipping via API with DroFx.pptxZuja dropshipping via API with DroFx.pptx
Zuja dropshipping via API with DroFx.pptxolyaivanovalion
 
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% SecureCall me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% SecurePooja Nehwal
 
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779Best VIP Call Girls Noida Sector 22 Call Me: 8448380779
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779Delhi Call girls
 
Mature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptxMature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptxolyaivanovalion
 

Dernier (20)

(NEHA) Call Girls Katra Call Now 8617697112 Katra Escorts 24x7
(NEHA) Call Girls Katra Call Now 8617697112 Katra Escorts 24x7(NEHA) Call Girls Katra Call Now 8617697112 Katra Escorts 24x7
(NEHA) Call Girls Katra Call Now 8617697112 Katra Escorts 24x7
 
Accredited-Transport-Cooperatives-Jan-2021-Web.pdf
Accredited-Transport-Cooperatives-Jan-2021-Web.pdfAccredited-Transport-Cooperatives-Jan-2021-Web.pdf
Accredited-Transport-Cooperatives-Jan-2021-Web.pdf
 
April 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's AnalysisApril 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's Analysis
 
Chintamani Call Girls: 🍓 7737669865 🍓 High Profile Model Escorts | Bangalore ...
Chintamani Call Girls: 🍓 7737669865 🍓 High Profile Model Escorts | Bangalore ...Chintamani Call Girls: 🍓 7737669865 🍓 High Profile Model Escorts | Bangalore ...
Chintamani Call Girls: 🍓 7737669865 🍓 High Profile Model Escorts | Bangalore ...
 
VidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptxVidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptx
 
Log Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptxLog Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptx
 
CHEAP Call Girls in Saket (-DELHI )🔝 9953056974🔝(=)/CALL GIRLS SERVICE
CHEAP Call Girls in Saket (-DELHI )🔝 9953056974🔝(=)/CALL GIRLS SERVICECHEAP Call Girls in Saket (-DELHI )🔝 9953056974🔝(=)/CALL GIRLS SERVICE
CHEAP Call Girls in Saket (-DELHI )🔝 9953056974🔝(=)/CALL GIRLS SERVICE
 
Invezz.com - Grow your wealth with trading signals
Invezz.com - Grow your wealth with trading signalsInvezz.com - Grow your wealth with trading signals
Invezz.com - Grow your wealth with trading signals
 
VIP Model Call Girls Hinjewadi ( Pune ) Call ON 8005736733 Starting From 5K t...
VIP Model Call Girls Hinjewadi ( Pune ) Call ON 8005736733 Starting From 5K t...VIP Model Call Girls Hinjewadi ( Pune ) Call ON 8005736733 Starting From 5K t...
VIP Model Call Girls Hinjewadi ( Pune ) Call ON 8005736733 Starting From 5K t...
 
FESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdfFESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdf
 
100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx
 
Data-Analysis for Chicago Crime Data 2023
Data-Analysis for Chicago Crime Data  2023Data-Analysis for Chicago Crime Data  2023
Data-Analysis for Chicago Crime Data 2023
 
CebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptxCebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptx
 
Week-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionWeek-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interaction
 
BigBuy dropshipping via API with DroFx.pptx
BigBuy dropshipping via API with DroFx.pptxBigBuy dropshipping via API with DroFx.pptx
BigBuy dropshipping via API with DroFx.pptx
 
Zuja dropshipping via API with DroFx.pptx
Zuja dropshipping via API with DroFx.pptxZuja dropshipping via API with DroFx.pptx
Zuja dropshipping via API with DroFx.pptx
 
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% SecureCall me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
 
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779Best VIP Call Girls Noida Sector 22 Call Me: 8448380779
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779
 
Mature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptxMature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptx
 
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in Kishangarh
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in  KishangarhDelhi 99530 vip 56974 Genuine Escort Service Call Girls in  Kishangarh
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in Kishangarh
 

Sanath pabba hadoop resume 1.0

  • 1. Sanath Pabba Mobile: +1-4168939098 Email: sanath.sanny@gmail.com Skype id : live:sanath.sanny_1 Synopsis  Engineering Professional with 5+ years of experience in Development, Maintenance and Production support in Bigdata technologies.  WorkingdirectlywiththeClientsregardingcritical requirementsandindependently workedonManyapps and provided many ideas to improve the system efficiency.  Written 12 automated scripts that optimized the duties and saved 70-80 hours of efforts per month.  Out of 12 automations,4 automationsgiven8-10% of data improvementswhichhelpsdatapreparation teams effort.  Expertise with the tools in Hadoop Ecosystem including Pig, Hive, HDFS, MapReduce, Sqoop, Storm, Spark, Kafka, Yarn, Oozie, and Zookeeper.  Excellent knowledge on Hadoop Ecosystemssuch as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.  Experience indesigninganddevelopingapplicationsin SparkusingScala to compare the performance of Spark with Hive and SQL/Oracle.  Good exposure with Agile software development process.  Experience in manipulating/analysing large datasets and finding patterns and insights within structured and unstructured data.  Strong experience on Hadoop distributions like Cloudera, MapR and HortonWorks.  Good understandingof NoSQLdatabasesandhandsonworkexperience inwritingapplicationsonNoSQL databases like HBase, and MongoDB.  Experienced in writing complex MapReduce programs that work with different file formats like Text, Sequence, Xml, parquet, and Avro.  Experience in Oozie and workflow scheduler to manage Hadoop jobs by Direct Acyclic Graph (DAG) of actions with control flows.  Experience in migrating the data using Sqoop from HDFS to Relational DBMS and vice-versa.  Extensive Experience onimportandexportdatausingstreamprocessingplatformslikeFlume andKafka.  Very good experience in complete project life cycle (design, development,testing and implementation) of Client Server and Web applications.  Strong knowledge on Data Warehousing ETL concepts using Informatica Power Center, OLAP, OLTP.  Having expertise knowledge on banking, insurance, and manufacturing domains.  Excellentproblem-solvingskillswithastrong technical backgroundand resultorientedteamplayerwith excellent communication and interpersonal skills.  Worked on creating the RDDs, DFs for the required input data and performed the data transformations using Spark-core.  Hands-on experience in Scala programming and Spark components like Spark-core and Spark-SQL. Technical Skill Set  BigData Technologies:HDFS,MapReduce,Hive,Pig,Sqoop,Flume,Spark,Kafka,ZookeeperandOozie  Languages: C, Java,Scala,Python,SQL,PL/SQL, PigLatin,HiveQL, JavaScript,Shell Scripting  DevelopmentMethodologies:Agile/Scrum, Waterfall VersionControlToolsGit,SVN,Bitbucket  RDBMS: Oracle,SQL Server  BuildTools: Jenkins,Maven,ANT
  • 2.  BusinessIntelligence Tools:Tableau,Splunk,QlikView,Alteryx  Tools: IntelliJIDE  CloudEnvironment:AWS  Scripting: Unix shell scripting,Pythonscripting  Scheduling:Maestro Career Highlights ProficiencyForte  Extensivelyworkedondataextraction,Transformationandloadingdatafromvarioussourceslike DB2, Oracle and Flat files.  StrongskillsinData RequirementAnalysisandDataMappingfor ETL processes.  Well versedindevelopingthe SQLqueries,unions,andmultiple table joins.  Well versedwithUNIXCommands&able towrite shellscriptsanddevelopedfew scriptsto reduce the manual interventionaspartof JobMonitoringAutomationProcess. Work experience  Sep2019 till currentdate in WalmartCanada as CustomerExperienceSpecialist(Part-time).  Jan 2018 to Mar 2019 in Infosys Limitedas SeniorSystemEngineer.  Apr 2015 to Jan2018 in NTT Data Global DeliveryServices asApplicationsoftware dev.Consultant. Project Details Company : InfosysLimited. Project : Enterprise businesssolution(Metlife Inc.) Environment : Hadoop, Spark,Spark SQL, Scala,SQL Server, shell scripting. Scope: EBS is a project where Informatica pulls data from SFDC and sends to Big Data at RDZ. Big Data kicksitsprocesswhenthe triggerfile,control fileanddatafilesare received.Allthe filescheckvalidations. Afterall the transformationsare done,the data is storedin hive,pointingtoHDFS locations.The data is synced to bigsql and down streaming process is done by QlikView team. Roles& Responsibilities:  Writing Sqoop Jobs that loads data from DBMS to Hadoop environments.  Preparedcode thatinvokessparkscriptsinScalacode thatinvolvesinDataloads,pre-validations,data preparation and post validations.  Prepared automation scripts using shell scripting that fetches the data utilizations across the cluster and notifies admins for every hour that helps admin team to avoid regular monitoring checks. Company : InfosysLimited. Project : BluePrism(MetlifeInc.) Environment : Spark,Spark SQL, Scala,Sqoop,SQL Server,shell scripting. Scope: BluePrism is a Source application with SQL Server as its Database. Big Data will extract the data fromBluePrism Environmentsmergetwosourcesintooneandloadthe dataintoHive Database.BigData also archives the data into corresponding history tables either monthly or ad-hoc basis basedon trigger file receivedfromBluePrism.ThisisaweeklyextractfromSQLserverusingSqoopandthenloadthe data through Scala. Jobs have been scheduled in Maestro. Roles& Responsibilities:  Prepared data loading scripts using shell scripting which invokes Sqoop jobs.  Implemented data merging functionality which pulls data from various environments.  Developed scripts that backup data using AVRO technique.
  • 3. Company : Infosys Limited. Project : Gross ProcessingMarginreports(Massmutual) Environment : Spark,Spark SQL, Scala,Sqoop,SQL Server,shell scripting. Scope:In GPMReports,we receive the input.csvfromthe Business. Basedonthe clientrequestwe need to generate 6 reports.We will receive the triggerfile anddata file for each report.Using shell script,we will perform validationon the trigger files, input file and paths that representation in Linux and HDFS, if everyvalidationsuccessfulandinvoke the hive scripttogenerate the outputfile andplacedinLinux.We will append the data in hive tables based on the output file. Migrating the Pig Scripts into spark scripts using Scala and report generation will be taken place by the spark and stores in Linux directory. Roles& Responsibilities:  Based on the business need, used to prepare the data using hive QL and Spark RDD’s.  By using Hadoop, we load the data, prepare the data, implement filters to remove unwanted and uncertain fields and merging all 6 reports from various teams.  Implemented8pre-validationrulesand7 postvalidationruleswhichinvolvesindata count,required fields and needful changes and in post validations we move the data to HDFS archive path. Company : Ntt Data Global deliveryservices. Project : Compliance Apps(National Lifegroup) Environment : Spark,Spark SQL, Scala,SQL Server,shell scripting,pig. Scope:Compliance appisagroupof nine adminsystems(AnnuityHost,ERL,MRPS,CDI,Smartapp,WMA, VRPS,PMACS,SBR).The processistoload the datafilesbasedonthe triggerfileswe receivedforthe nine admin systems in to HDFS. There are three types of Load that takes place. They are: 1. Base load 2. Full load 3. Deltaload Tobuildthe workflowtoloadthe datafilesintoHDFSlocationsandintohive tables.We needtocreate hive tableswithoptimizedcompressedformatandload the data into the tables.To write hive script for full load and write the shell script to create a workflow. We use Pig/spark for the delta loads and shell script to invoke the hive for the full load/Historyprocessing.Then,schedule the jobsinMaestro for the daily run. Initially for delta load, we were using Pig scripts. Company : NTT Data Global DeliveryServices. Project : ManufacturingCostWalk Analysis(Honeywell) Environment : Sqoop,Shell scripting,Hive. Scope: The Manufacturingcost walkapplicationusedtostore the informationaboutthe productswhich are beenmanufacturedbyHoneywell.Theyusedtostore the datainSharePointlistsonweeklybasis.But still it is very difficult for them to handle the data using share point because of long time processing. So we proposeda solutionforthemwithHive andSqoop.But theirsource of file generationisfromcsvand xlsfiles.So,we have startedimporting data into hive and processing data based on their requirement. Academia  Completed Post graduation from Loyalist college in Project Management (Sep 19- May 20)  Completed diploma from Indian institute of Information Technology (Hyderabad) in Artificial Intelligence and data visualization (April 2019 – Aug 2019)  Completed Graduation Under JNTU-HYD in Electronics and Communications (Aug 11 – May 14)