1. Tarun Poladi
MSBI Data Engineer
+1(909) 713-4720
tarun.sqldev04@gmail.com
https://www.linkedin.com/in/tarunraosqldev
Summary
• Analytics Professional with a track record of deploying Business intelligence tools & Data Engineering
methodologies for leading Data-driven research to provide market insights & deploy user-centric solutions.
Proven expertise in generating insightful reports to deliver compelling business value to key stakeholders and
clients.
• Having Around 7+ years of IT experience in Database Development of SQL Server and Oracle and providing
Business Intelligence solutions using Microsoft BI Stack (SSIS, SSRS, SSAS-Multidimensional Cube, Tabular
Models, Power BI, Excel-Power Pivot, Power Query) of SQL Server.
• Experienced in Star Schema and Snowflake Data Modeling Techniques as well as following Bill Inmon and Ralph
Kimball principles in Managing Data marts and EDW Environments.
• Excellent experience in writing T SQL Queries using various objects like views, tables, Joins, Sub Queries, triggers,
CTE’s, stored procedures, and other advanced SQL concepts with query optimization techniques using the
Partitioning and indexing concepts.
• Experience in using Data bricks for handling all analytical process from ETL to all data modeling by leveraging
familiar tools, languages, and skills, via interactive notebooks or APIs.
• Worked on large and disparate datasets (including CSV’s, JSON Files, Paraquet) by using Pyspark, NumPy and
pandas.
• Imported the data from different sources like AWS S3, Local file system into Spark RDD.
• Experience with developing and maintaining Applications written for Amazon Simple Storage, AWS Elastic Map
Reduce, and AWS Cloud Formation and services such as EC2, S3, CloudFormation, RedShift, and Dynamo DB
which provides fast and efficient processing of Big Data.
• Exploring with the Spark improving the performance and optimization of the existing algorithms in Hadoop using
Spark Context, Spark-SQL, Data Frame, Pair RDD's, Spark YARN.
• Developed user Friendly Rich PowerBI Reports / dashboards with intuition to UX Design using Drill through,
Bookmarks and Integrating R ggplot2, plotly, python Seaborn, scatterplots, Heat maps, Linear Regression,
Correlation plots Leaflet Maps into PowerBI.
• Good experience working on Hybrid Tables and composite models in Power BI as well as Analytic data warehouses
like snowflake and Azure synapse.
Technical Skills
Big Data Eco System Hadoop, MapReduce, Pig, Hive, YARN, Kafka, Zookeeper, Spark, Elastic Search, Splunk, Parquet .
Languages T-SQL, Python, R, C#, Shell Scripting
Databases MSSQLServer, PostgreSQL, Azure Blob Storage, Apache Cassandra, Azure Data lake, Mongo DB
IDE/ Programming Tools Visual Studio Code, PyCharm, SSMS, Azure Data Studio,
Jupyter Notebooks
Operating Systems Windows, Unix, Linux.
CI/ CD Terraform, Jenkins, Azure Devops, TFS
Web Technologies HTML, CSS, XML, Java script, jQuery, Bootstrap
Libraries and Tools Pyspark, Pandas, NumPy, Boto3, SSIS, SSRS, PowerBI,
Jira, Scrum, Agile Methodologies
2. 2
Professional Experiences
Paychex, West Palm Beach-FL Sept’ 2020 – Present
Data Engineer
• Worked extensively on AWS Components such as Elastic Map Reduce (EMR), Athena.
• Developed SQOOP scripts to migrate data from Oracle to Bigdata Environment.
• Created data partitions on large data sets in S3 and DDL on partitioned data.
• Converted all Hadoop jobs to run in EMR by configuring the cluster according to the data size.
• Extensively used Stash, Git-Bucket for Code Control.
• Monitor and Troubleshoot Hadoop jobs using Yarn Resource Manager and EMR job logs using Splunk and
Kibana.
• Created data pipeline for different events of ingestion, aggregation, and load consumer response data in AWS S3
bucket into Hive external tables in HDFS location to serve as feed for PowerBI dashboards.
• Worked with different file formats like JSON, AVRO and parquet and compression techniques like snappy.
• Developed shell scripts for dynamic partitions adding to hive stage table, verifying Json schema change of source
files, and verifying duplicate files in source location.
• Created Athena external tables using Hive DDL to point to the data stored in AWS S3.
• Built data lake ecosystem using AWS technologies, such as S3, Athena, EMR, to ingest and process Kinesis data
Streams
• Converted Hive queries into Spark transformations using Spark RDDs.
• Exploring with the Spark improving the performance and optimization of the existing algorithms in Hadoop using
Spark Context, Spark-SQL, Data Frame, Pair RDD's, Spark YARN.
• Developed Preprocessing job using Spark Data frames to flatten Json documents to flat file.
• Load D-Stream data into Spark RDD and do in memory data Computation to generate Output response.
• Experienced in writing live Real-time Processing and core jobs using Spark Streaming with Kafka as a data pipe-
line system.
• Worked extensively with importing metadata into Hive using Python and migrated existing tables and applications
to work on AWS cloud (S3).
• Web applications leveraging the data into a real-time pipeline using AWS Kinesis, Mongo, Apache Spark,
Elasticsearch, AWS S3.
Environment: Pyspark, Python, Databricks, Azure, AWS, Snowflake, AWS Glue, Splunk, PowerBI, SQL, Terraform,
Bash Scripting
Oasis Advantage Payroll Services, West Palm Beach-FL Oct’ 2018 – Aug’ 2020
Data Engineer
• Involved in designing new Modern Enterprise Data Warehouse Architecture using Azure platform and using Azure
Analysis Services as Ad-hoc reporting Solution and Data visualizations through PowerBI on-Premises Report Server
as well as PowerBI Cloud Services.
• Integrate various legacy data Sources like legacy unidata into SQL normalized Form for clean instances of Data
elements, Oracle database, third party data through web services into one centralized Data warehouse in Azure
SQL Datawarehouse using ETL Techniques and designed SSIS Frame works along with data architect for
continuous flow of data.
• Created PowerBI Dashboard sourcing the data from Azure Cosmos DB and to secure it with hierarchal Row Level
Security for the end users to consume the interactive PowerBI Reports as well as dashboards.
• Have Successfully Migrated Existing Izenda canned Reports, to PowerBI Reports in On-Premises Report Services
for business users to enhance the usability of reports.
• Implemented CDC using OUTPUT Clause in SQL.
• Worked on Performing sentiment analysis for Customer Surveys received through Qualtrics in PowerBI Using NPS
score metrics with help of Azure Cognitive services.
3. 3
• Monitoring the Power BI On-Premises Report Services, Cloud versions. Usage Statistics and Refresh schedule
runtimes in PowerBI Folders, and Workspaces respectively.
• Upgrading the PowerBI report server for every 4 months whenever Microsoft releases the new version for report
services by using report server backup mechanisms.
• Converting existing Pivot report Mockups in to interactive PowerBI reports with all the latest PowerBi updates
and Paginated reports Sourcing the data from SQL server as well as web Api’s and Scheduling them for mail
delivery daily Morning.
• Analyzing the key business needs and incorporating Predictive Analytics like K-Means Clustering algorithms to
predict the financial sales for sales department.
• Created PowerBI Dashboards Consuming Azure Cosmos DB and Converting the unstructured data into
Consumable way to present it to the end users.
• Worked on fixing Bugs in the existing SSIS Packages and made improvements to get down the run time of the SSIS
Package.
Environment: T-SQL, PowerBI, SSRS, SSIS, Visual Studio, Azure SQL DB, PowerShell.
Tennessee Valley Authority (TVA), Chattanooga-TN Nov’ 2017 – Sept’ 2018
SQL PowerBI/R-Shiny Developer
• Have created new data mapping document for implementing a DataMart to get into the analysis server for
reporting as well as Ad-hoc reporting purposes from Power BI.
• Had been a team member in creating data governance for PowerBI cloud solutions across the TVA enterprise wide
collaboration with business users.
• Have Successfully Migrated Existing Actuate Reports, Cognos reports to PowerBI Reports in Premium workspace
for business users to enhance the usability of reports.
• Lead the PowerBI training to business users where users can connect to Tabular Analysis Services and can create
their Reporting for future Use.
• Have Worked on Statistical Analysis Reports which involves Performing Hypothesis testing of data samples in
Microsoft SQl server-2016 with R (in-built) services and had created the reports Power BI Service with R visuals
• Created PowerBI Reports which involves Linear regression Models (Predictive models) for Data from Web API
integrations
• Enhanced reporting capabilities & created real-time dashboards pertaining to KPI metrics for senior leadership
and have Embed PowerBI dashboards into MVC model with token Authorization for larger audience viewing.
• Power BI reports were creating using dynamic URL Filtering in On-Premises Report Server where some of the
secure data was not allowed to move to Cloud workspace.
• Designed parallel jobs via transformers, Aggregator, Filter, Join, Lookup, Shared container, Funnel, Pivot, etc.
• Created Complex ETL Operations using script components and also using stored procedures in process of getting
legacy data into new database by creating logic to validate the data before loading to the new tables and delete
the duplicate data using SQL Server Integration Services tool (SSIS).
• Created/Modified Stored Procedures, functions, Triggers and Power shell commands as per the requirement.
• Created Sub-Reports, Drilldown-Reports, Summary Reports, and Parameterized Reports in using PowerBI and
maintained them in PowerBI report Server.
• Created some complex DAX expressions in creating calculated columns and Measures for the Power BI
Dashboards to transform the unstructured data in needed form for the business users to make use of them.
Environment: SQL Server 2014, MSBI Tools, Pandas, Tabular Model, R, Team Foundation Server, Power BI Report
Server, Power BI Premium Workspaces, Server Management Studio, Microsoft Azure, Actuate Reports
4. 4
PSECU Bank, Harrisburg- PA Aug’ 2016 – Oct’ 2017
MSBI Developer
• Involved in planning, Defining and Designing information focused around business necessities and gave
documentation
• Responsible for writing T-SQL queries, Dynamic SQL, complex stored procedures, functions, triggers, Views.
• Created Database objects like tables, Stored Procedures, Views, Triggers, Rules, Defaults, Client characterized
information sorts and capacities in SQL server.
• Worked on migrating 200 SSIS Packages from 2008R2 to 2014 Server.
• Designed SSIS packages to concentrate, exchange, load (ETL) existing information into SQL Server
• Performance tuned the existing stored procedures to reduce the existing Query times.
• Creating and maintaining PowerBI Reports and Dashboards and embedding them inside on premises
Organizational apps for the larger audience to view them.
• Worked on Creating Dynamic Parameterized Power BI reports to filter the Data from One report To the Other.
• Worked on Paginated reports in PowerBi Cloud and scheduled them to mail in form of PDF reports for End users
Consumption.
• Worked on SSIS Packages in cleansing unstructured And Missing Data from Legacy Systems to new EDW
Environment in Package Deployment Model.
• Performed the Slowly Changing Dimensions Using Hash Byte Techniques for Better Performance of SSIS Packages
and for parallel Loading of Data.
Environment: SQL Server 2021, SSIS, SSRS Tools, SSAS Multi-Dimensional Model, Power BI Pro Workspaces, Service
now.
IT Commence, Irving-TX Nov’ 2014 – July’ 2016
SQL/BI Developer
• Database development experience with Microsoft SQL Server in OLTP/OLAP environments using integration
services (SSIS) for ETL (Extraction, Transformation and Loading).
• Designed SSIS packages to bring data from existing OLTP databases to the new data warehouse by performing
different kinds of transformations like lookup transformations, merge joins, fuzzy look-ups, derived columns,
merge join, conditional split, data conversion with multiple data flow tasks.
• Created and implemented complex SSIS package to migrate data from multiple data sources for data analyzing,
deploying and dynamic configuring of SSIS packages.
• Implemented Event Handlers and Error Handling in SSIS packages and notified process results to various user
communities.
• Maintain and optimize the existing databases; monitor database performance and growth, performance tuning,
generate reports according to the business requirements using SQL Server Reporting Service (SSRS) and populate
data warehouse using SQL Server Integration Service (SSIS).
Environment: SQL Server 2008/2021, SSIS, SSRS Tools, Excel, C#
Education, Certifications and Coursework
• Master in CS at University of Houston-Clearlake (UHCL), Houston-TX.
• Bachelor’s in Computer Science and Engineering (CSE) at Jawaharlal Nehru Technological University, India.
• DP-203 Azure Data Engineer
• MCSA (Microsoft Certified Solutions Associate): BI Reporting
➢ 70-778 Analyzing and visualizing data With PowerBI
➢ 70-779 Analyzing and visualizing data With Excel