1. RESUME
Parthiban Ranganathan
Thousand Oaks, CA.
Phone: +1 8059907090
PROFILE SUMMARY
• Over Seven (7 years 3 months) years of IT experience remarkably in Data Warehousing technology,
Business Intelligence and a Big Data aspirant with knowledge in Hadoop.
• Have worked in multiple assignments in Healthcare [Client: Anthem Inc. - Membership, Claims and
Provider Subject areas of one of the largest Healthcare Data Warehouse EDWARD] and Manufacturing
[Client: Cummins Inc.] domains.
• Have solid understanding of OLAP concepts, OLTP Conepts, Data warehousing concepts.
• Possess strong knowledge of Entity-Relationship concept, Facts and dimensions tables, slowly changing
dimensions and Dimensional Modeling (Star Schema and Snow Flake Schema).
• Experience in Data Integration and ETL with various data sources like Teradata, Oracle, DB2, Sybase,
SQL server and non-relational sources like flat files(CSV, Binary, COBOL VSAM, XML ) into staging area.
• Expertize in SQL Scripting.
• Experience in Business Intelligence like Business Modeling and Reports creation, Scheduling &
Publishing.
• Possess excellent analytical, problem solving, communication and interpersonal skills, with ability to
interact with individuals at all levels and work as a part of a team as well as independently.
• Capable of grasping new technology, process and tools quickly and mentor of novice.
TECHNICAL EXPERTIZE
Type Particulars
Database Teradata, Oracle, MS SQL Server, DB2, Sybase
ETL Informatica Power Center, Teradata BTEQ, MS SQL Server TSQL
Stored Procedure, Oracle PL/SQL, SSIS
BI Tools Business Objects, OBIEE, Cognos, SSRS
Data Analysis SSAS, MS EXCEL
Scripting UNIX Shell Scripting, PERL
Data Modelling MS VISIO, ERWIN
Versioning Clearcase
Scheduling Tool WLM, Control M, Appworx, DAC
Other Tools RequestPro, JIRA, ClearQuest
Database Utilities Teradata MLOAD, FLOAD, FASTEXPORT scripts, TPUMPs,
Oracle PL/SQL cursors,triggers, stored procedures, functions,
packages
Database Tool MS Management Studio, Teradata SQL Assistant,
Oracle – Toad, SQL Developer and SQL*PLUS.
Secondary Programming Skills C, C++, HTML, EXCEL MACRO.
Operating Systems Operating Systems UNIX, Windows 2000/2003/XP/7/8/10.
2. PROFESSIONAL EXPERIENCE
Designation Company Period
Associate Cognizant Technology Solutions Nov 2011 – Till date
Programmer Analyst Cognizant Technology Solutions Nov 2009 – Nov 2011
EDUCATION
Degree and Stream College/University Year of Passing
Bachelor of Engineering in
Electronics and Communication
Institute of Road and Transport Technology,
Erode.(Anna University Affiliated)
2008
PROJECTS AND RESPONSIBILITY
Duration Jan 2014 – Dec 2016
Project Title Anthem Provider Finder & Procurement Data Mart
Projects
Role ETL and BI Tech Lead
Funcational Area Healthcare & Data Warehousing, ETL, Business Intelligence
Client Anthem Inc.
Operating Systems Windows XP, UNIX
Tools Informatica, UNIX, Shell Scripting, MS SQL Server 2008/2014, T-SQL,
BCP, Teradata BTEQs, MLOAD, FLOAD scripts, WLM, ClearQuest,
RequestPro, JIRA, Business Objects, Infogix, MSBI (SSAS, SSRS).
PROJECT DESCRIPTION:
• Provider Finder is Find a Doctor tool in WellPoint consumer portal to find Doctors, Facilities,
Pharmacy. It is a critical application for Anthem’s business. Multiple migration projects and multiple
enhancement projects.
• To build Procurement Data Mart that will enable WellPoint to create enterprise-wide vendor reporting
using multiple data sources (SourcePoint, Plan view, Fieldglass, SharePoint, D&B, etc.). Data Mart will
deliver reporting data via customized reports and dashboards to Global Vendor Management, IT
Vendor Management, and Business Process Outsource Risk Committee stakeholders. Data Mart will
also cater to the Procurement’s audit findings.
RESPONSIBILITIES:
• Worked as an Tech Lead and involved in activities like Requirement Gathering & Analysis, Estimate
Project Cost, Design, Coding & Development, Unit testing, System Testing , User Acceptance Testing
Support, Implementation Support, Maintenance Support, Checkout Support, Warranty Support.
• Responsible for understanding, analyzing the Requirement, interacting with SAs and Business
stakeholders for clarification and interpreting the the requirements into coding..
• Responsible for presenting the Design to Clients and getting them signed off.
• Involved in Data Model changes, working along with the Project Data Architect.
• Responsible for Creating LOEs in Initiation Phase, Project budget reports and presented to Project
Management.
• Active member of Right In-Time team, responsible for Analyzing, Resolving real time issues in the
Production, Bug fixing and taking load related decisions.
• Experience in creating Business and Technical Requirements Documentation.
3. • Created High Level Design and Detailed Design, Technical Specification, Attribute Mapping, Logical
Data Models and Physical Data models, Data flow and HLD and LLD diagrams with MS VISIO in the
Design phase.
• UNIX Shell scripts for File Processing, File transfers, Archiving, Auditing, BO report scheduling and
error handling, Infogix Controls.
• Created Jobs/Jobsets for triggering Informatica Workflows and Shell Scripts in WLM/ Scheduling
Tables/Jobs in Control-M.
• Tuned mappings using Power Center-Designer and changed logic of existing code to provide
maximum efficiency and performance.Tweaked SQL queries in the Stored Procedures to improve
performance and tuned Teradata SQL queries to solves the spool space issues.
• Created Use cases, Test Scenarios, and Test Results/Reports for Unit Testing and Provided Test Data
for System and User Testing. Experience in Clearquest, JIRA for Creating, Triaging, Deferring, Closing
and Canceling Defects.
• Suported DBAs, Informatica Admins, UNIX Admins and WLM/Control-M Admins, BO Admins in
Migration/Deployment of Codes/Changes and created Migration documents for all environments.
• Responsible for in Versioning the Informatica Codes in XML format and other Codes with Clearcase.
• Experience Implementation Support, Business checkout, Technical Checkout, Post Production
Validation and Testing Support.
• Responsible for providing, Presenting Project RAG Status reports to the Clients.
• Responsible for Implementing all the Project changes handled by my Team in Monthly release
activities, on average 6 to 7 Projects/SSCRs everymonth starting from Sep 2014 to Dec 2016.
ACHIEVEMENTS:
• Successful migration of Provider data from EPDSV2 to EDWARD which is built on Teradata.
Write !complex BTEQ scripts and use of MLOAD and FLOAD utilities to load Provider data from
EPDSV2.
• Automated data validation (Reconciliation,comparing previous/current eligible data loaded etc..)
involved in the each and every Production Load Steps with around 2500 validation queries and
created BO reports with validations data and scheduled them to Run appropriately and automatically
send mails to decision making with PASS/FAIL status with reports attached.
• Successfully delivered Data migration projects, changing the sourcing of data from Legacy systems
to EPDSV2 into Provider Finder data layer.
• Converted Weekly loads to Daily loads by changing the Kill-Fill load strategy to Change Data Capture
in the Landing Zone and Stage Layers, thus improving the Functionality of the System.
• Decoupled Stage layer and Target layer loads eliminating the Dependencies between load steps
which involved 14 Source systems (WGS, SSB-CA,SSB-IN, CS90, FACETS, ACES, WEST, Pharmacy,
Dental, Vision, VA, BCBSA, CACTUS, GA) increasing the Maintainability of the system in case of
Invalid Data loads or errors in the data load.
• Spider Application took 4 hours to load 1 GB of XML Source file into the SQL Server as early as
possible, as the data in the DB is critical for business. Reduced the 4 hours running time to 45 mins,
by removing the File dependency, tweaking the SQL codes and configuration changes in the
Informatica Sessions and Workflows.
4. Duration Dec 2012 – Dec 2013
Project Title Cummins Enhancements and Projects
Role Senior ETL and BI Developer
Funcational Area Manufacturing & Data Warehousing, ETL, Business Intelligence
Client Cummins Inc.
Operating Systems Windows XP, UNIX
Tools Informatica, UNIX Shell Scripting, Oracle 11g PL/SQL, Appworx & DAC
Scheduler, OBIEE 11g, Clearcase, MSBI (SSIS, SSAS, SSRS).
PROJECT DESCRIPTION:
• To capture the bugs in the Cummins’ applications such as EBU (engine business unit), Corp Apps
(Corporate Application) and modification of the application in order to fix the existing bugs and
adding more functionality and capability. !The changes to the existing ETL Codes, Database etc. are
designed and developed based on the Change Requests called CRQs. Delivered many CRQs
notably, SSA, FLD, RPV, NAFTA BOM Correction. NAFTA (North American Free Trade Agreement)
application creates reports like Billing of Materials, Shipment details and Engine Component Items.
Building of new functionality to ensure the correctness of the reports created in the NAFTA
application. To Design, Build new Error Mechanism and correction of reports in MacTrack (Material
Cost Tracking) Application.
RESPONSIBILITIES:
• Worked as an Sr. Informatica and PL/SQL, BI Developer and involved in activities including
Requirement Analysis, Design, Coding & Development (involves policy rule coding, sequence flow,
status flow, and dictionary entries), Unit testing, System Testing, UAT Support, Implementation,
Maintenance.
• Record data errors from the Sources of MacTrack Application and creates reports out of it for the
Clients. Building Error Mechanism for the MacTrack System and send reports to the Business for
making decisions on loading the data. Error Mechanism using Informatica Mapping/Workflow and
publish report to the users in the format of CSV Files.
• Developed Informatica Mappings with Reusable Mapplets,Transformations (Joiner, Sorter,
Aggregator, Expression, Lookup, Router, Filter, Update Strategy, Sequence Generator, Normalizer
and Rank), Worklows with Reusable Tasks (Sessions, Command, Email) and Non-Reusable Tasks
(Decision, Event Wait, Event Raise, Timer, Assignment, Worklet, Control).
• Developed Dashboards/Reports in OBIEE, Reports in SSRS. Analyzed and corrected the errors in
the Bill of Materials (BOM) reports of Cummins Power Generation Unit of North American Free Trade
Agreement (NAFTA) regions. Correction of OBIEE dashboard for BOM.
• Added a column Part Number in the Business model of the BOM, in the Repository and use that in
the aggregation of the BOM list report.
Duration Sep 2009 – Dec 2012
5. Project Title Anthem EDWARD Development Projects
Role Informatica & Teradata Developer
Funcational Area Healthcare & Data Warehousing, ETL
Client Anthem Inc. (then Wellpoint Inc.)
Operating Systems Windows XP, UNIX
Tools Informatica Power Center 8.6.1, Teradata, UNIX Shell Scripting, Putty,
Clearcase.
PROJECT DESCRIPTION:
• EDWARD (Enterprise Data Warehouse and Reporting Depot) is a Huge Data Warehouse with data
integrated from all the source systems of Wellpoint such as CS90, NASCO, ACES, WGS, FACETS
etc. and all Subject areas such as Membership, Claims, Provider etc. The target layer is itself 250+
tables with huge amount of data, and thus there were hundreds of code components including
Informatica Workflows, Teradata BTEQs etc.This assignment is to perform Functional testing of all
existing components; Modification of existing codes as per the standards and changes to the
existing Informatica Mappings focusing on the EDWARD improvement.
RESPONSIBILITIES:
• Worked as an Senior Informatica/Teradata Developer and involved in activities including
Requirement Analysis, Design, Coding & Development (involves policy rule coding, sequence flow,
status flow, and dictionary entries), Unit testing, System Testing, UAT Support, Implementation,
Maintenance.
• Created Technical Design documents.
• Developed Informatica codes for loading the files in the Landing Zone tables.
• Implemented fixes/solutions to the Bugs encountered during the functional testing and Regression
Testing support.
• Developed BTEQ scripts involving complex queries to derive benefits/coverage of a
member/spouse/dependent incase of multiple insurances subscribed by member, spouse from
other carrier and of dependents.
• Created Test Cases, Test Plans and Test results and documented them. Unit tested the codes
written to check the data loaded in the target is as per the logic and the requirement. Supported SIT,
UAT testing and Implementation of the project.
ACHIEVEMENTS:
• This is one of the biggest project I was invloved in writing 10K lines of more than 40 BTEQ Scripts to
achieve the Coordination of Benefits for the NASCO source system.
• Created SQL Parser with PERL Text Mining, for validating the BTEQ SQL Scripting standards and
cosmetics of the code.
• Pivoted Plan and corresponding Rates with various categories. Created mapping with Normalizer
and Aggregator Transformation for Pivoting Excel data, converting multiple rows into single row,
concatenating attributes in the Plan Spreadsheet and combining(Using Union Transformation) Rate
data with multiple categories like (Family with one child, with two child etc.) from another Excel sheet
• Created reusable Excel templates with checkbox to record test results of each component and
reduced the documentation time.
• Created SCD Type 2 Mappings using Lookup & Update Strategy Transformation to load plan,
pricing, member etc. data into the ODS.