AbInitio ABHINAV
AbInitio ABHINAV
AbInitio ABHINAV
AbInitio
Abhinav Pande,PMP Specialist Loc:
Pasadena,CA
Profile
Certifications
Education
Bachelors in Computer Engineering from State University of New York at Buffalo
(Deans List)
Abhinav Pande 1
Software Expertise
Datawarehousing Ab Initio (GDE 1.14.37
Co>Operating System 2.14.3/2.13.8/ 2.7.3)
Abhinav Pande 2
Key, Partition by round robin, reformat, rollup, join, scan, gather,
replicate, merge etc.
Extensively used the Ab Initio tool’s feature of Component, Data and
Pipeline parallelism.
Used Teradata Utilities Fast Load, Multi-Load, and Tpump for data
loading.
Configured the source and target database connections using .dbc files
Generated DB configuration files (.dml, .cfg) for source and target tables
using db_config and modified them according to the requirements.
Used sandbox parameters to check in and check out of graphs from
repository Systems.
Developed various Ab Initio Graphs for data cleansing using function like
is_valid, is_defined, is_error, is_defined, string substring,
string_concat and other string_* functions.
Created Sub Graphs to impose application/business restrictions.
Developed UNIX Shell Scripts to automate file manipulation and data
loading.
The Enterprise Data Warehouse provides integrated information and services that
enable our Lines of Business partners to attract new customers, retain existing
customers, make informed financial decisions, and improve business practices.
The EDW is comprised of a series of integrated dimensional data marts. These data
marts enable end-users to directly query historical data relating to JP Morgan Chase
accounts, transactions, households, customers, products, and prospective customers.
Worked for the Enterprise Data Warehouse Team of JP Morgan Chase in Columbus as
a Developer / Analyst
Gathering business requirement from business users for the financial interfaces
coming from various source systems of JPMorgan Chase and BankOne.
Designing and developing the ETL processes to process and load data into Enterprise
Data warehouse (EDW) coming from various source systems
Developed a lot of graphs in the ETL tool AbInitio for Deposit & Customer team of the
bank.
Worked for Production Support during evenings and weekends.
Used various components like sort, reformat, join, lookup, filter, dedup etc
Abhinav Pande 3
Various performance tuning techniques used to improve the performance
Responsible for the automation of Ab Initio graphs using Maestro Scheduling tool
Data warehouse is built to handle the immense amount of data produced in the
bank’s operations and allow that data to be analyzed for correlations that would
provide commercial advantages. This Data Warehouse on Oracle 8.1.7 database is
built by integrating data from the bank’s major data source systems like DB2 and Flat
file data and external data sources available on different platforms.
Analysed the Bank’s business processes and interacted with the end users for
informational requirements.
Involved in the design and implementation of the Data model for the Data
Warehouse using Star Schema.
Developed and supported the extraction, transformation and load process
(ETL) for the Data Warehouse from heterogeneous source systems using Ab
Initio.
Used most of the generally used Ab lnitio components like Reformat, Join,
Rollup, Normalize, Dedup, Sort, Input table, Output table, etc and
worked with parallelism using various partition methods like partition by
key, partition by expression, etc.
Developed source watchers that look for incoming flat files (Delta) from other
servers and once found the required flat file will create indicators for the down
streams that will indicate availability of the file.
12-way multi file system was implemented to partition the data and
various operations were run in parallel on the partitioned data.
Configured the source and target database connections using .dbc files
Created .xfr and .dml files for various transformations and specifying the
record format.
Incorporated data parallelism into graphs by using partition by key and
partition by Round Robin. Used partition by Round Robin to avoid skew.
Phasing was done to the complex graphs to avoid the table lock situations
while loading and updating the tables.
Abhinav Pande 4
Involved in Performance Tuning of Ab Initio graphs using Various Ab Initio
performance techniques and best practices such as using Lookup’s instead of
Joins and used in memory sort where ever possible.
Deployed, tested ran the graph as executable Korn shell scripts in the
application system.
Environment: Ab lnitio GDE 1.12.3, Ab lnitio Co>Op 2.12.3, Oracle 9i, AIX
5.0, Solaris 5.8, Harvest.
This project deals with the banking transactions at regional and Country level. They
follow two systems Advantage and E-Advantage. Advantage Application captures
data from Flat files and Excel Sheets, where as E-Advantage Application pumps
huge data into the production database, which makes it impossible to maintain
beyond 90 days. Each location at the end of the day has to transfer the data into
central zones and each location daily generates reports for different purposes like
Loan portfolio, Non sufficient fund port folio. They needed extensive Data
Warehousing to maintain historical data at a central location for integration and
analyze the business information in different locations according to the profit areas,
which could serve the purpose of a DSS management.
Abhinav Pande 5
Developed shell scripts to automate file manipulation and data loading.
Environment: Ab Initio Co>OS vs. 2.11.8, GDE 1.12.5.2, Oracle 9i, Toad
7.6.0.11, HP UX 11.x, IBM PCOM 4.2, TSO/ISPF, Korn Shell Scripting.
Client: Walgreens, IL
Project 5 June 2001 to May 2003
ETL Developer
Environment: Ab Initio Co>OS vs. 2.11.8, GDE 1.12.5.2, Oracle 9i, Toad
7.6.0.11, HP UX 11.x, IBM PCOM 4.2, TSO/ISPF, Korn Shell Scripting.
Abhinav Pande 6