JAFFER Hussain SYED Resume
JAFFER Hussain SYED Resume
JAFFER Hussain SYED Resume
872-806-4363
syedjaffer436@gmail.com
• Over 9 years of IT experience in analysis, design, development, data conversion, Data Quality, Data
Warehouse, Business Intelligence, Testing, documentation, and implementation of client/server applications
• Strong working experience in the Data Analysis, Design, Development, Implementation and Testing of Data
Warehousing, ODS and DataMart's using Data Conversions, Data Extraction, Data Quality, Data Transformation
and Data Loading (ETL).
• Experience in BI reporting tools using Business Objects Cognos, and Tableau.
• Knowledgeable in Informatica Cloud Data Quality (CDQ) for data profiling, cleansing, standardization, and
matching
• Experienced in Azure Data Factory for data integration, transformation, and orchestration.
• Interpreted logical and physical data models for business users to determine common data definitions and
establish referential integrity of the system.
• Experience architecting highly scalable, distributed systems using different open-source tools as well as
designing and optimizing large, multi-terabyte data warehouses.
• Able to integrate state-of-the-art ETL technologies into the overall architecture and lead a team of developers
through the construction, testing and implementation phase.
• Extensively used Informatica Data Quality, Informatica Repository manager, Power center Designer, Workflow
manager and Workflow Monitor.
• Familiar with data architecture including data ingestion pipeline design, ETL architecture, data modeling and
data mining, GCP Big Query
• Expert in design and development of complex ETL mappings and scripts using Informatica PowerCenter and
Informatica Data Quality.
• Hands on solid experience in Performance Tuning of source, target, mappings, transformations, and sessions
in Power Center and Data Quality.
• Hands on solid experience in using REST APIs, Salesforce Connector, parsing JSON and XML Data into relational
format using Informatica Data Quality.
• Experience in interfacing the data from legacy systems to Oracle applications using SQL*Loader. Experienced
in writing SQL statements and PL/SQL code in Oracle
• Experience in Building Business Objects Universes and using various Data Providers like Universes, Stored
Procedures, Free Hand SQL, Personal Data Files for retrieving data and creating Simple, Complex and Ad-hoc
Reports.
• Strong logical, analytical and communication skills with Solid business acumen. Ability to effectively gather and
refine report requirements from Users and Management
• Strong Team Management skills and excellent team player.
• Familiar with Elastic Search Engine and Python Scripting.
• Experience in a wide range of Business Domains –HealthCare, Financial, Government, Telecom, Banking, and
Insurance.
Technical Skills
Data Warehousing Informatica Data Quality & PowerCenter
5.1.2/6.x/7.1.4/8.x/9.x/10.5. (Source Analyzer,
Designer, Mapping Designer, Mapplet,
Transformations), Power Exchange, B2B Data
Transformation, Metadata Manager, GCP Big Query
Reporting, Bug Reporting & Version Control Tools Business Objects 5.0/5.1, Micro Strategy 7.5, Tableau
2020.4, Jira Align, Jira User Stories, SQL*LOADER,
Crystal Reports 8.5, SQL Navigator 4.5.1.53, TOAD,
SVN, Control-M, Tidal Enterprise Schedule, Devops
O/S Windows 95/98/NT/2000/XP/10/11, MS-DOS, Unix,
Linux, Sun Solaris.
Languages PL/SQL, SQL, Transact-SQL, UNIX shell scripting,
Python, Teradata-Bteq Scripts
Databases Oracle 9i/10g/11g/12c/19c, Teradata 12.0/14.0/15.0,
DB2, SQL Server 2014/2019, Sybase 11.5 and MS
Access97/2000, SAP
Data Analysis Data Profiling, Data Quality /IDQ 8.6.1/9.6.1/10.5
PROJECTS
WellSense is a nonprofit health plan that provides health insurance coverage to Massachusetts and New Hampshire
residents, including low-income, underserved, disabled and elderly populations, established by Boston Medical Center.
Generating Historical data Extracts from legacy and new warehouse in partnership with CMS, Medicare programs,
Matrix, and other Health solution partners
Responsibilities:
• Project involved usage of most of the transformations like Transaction Control, Active and Passive look up
transformation, Router, Filter, Expression, Aggregator, Joiner, Update Strategy, Stored Procedure,
Unstructured data transformation, SQL transformation and more
• Extensively used IDQ transformations in day-to-day work such as - Sorter, Filter, Expression, Consolidation,
Match, Exception, Association and Address validator transformations
• Created various IDQ scorecards for Data profiling, worked on deduping the Company wide data by using
Match and Merge IDQ process
• Developing data quality models to identify issues with the data and recommending solutions to improve the
data quality
• Worked with Azure Data Factory for integrating data from on-premises sources and cloud-based applications
• Experience development of SCD Type I and Type II with the help of MD5 hash function
• worked with IDQ on data quality for data cleansing, robust data, remove the unwanted data, correctness of
data, and Address verification using IDQ Address Doctor
• Experience working with extracting and loading data directly into Salesforce objects using Informatica
PowerCenter
• Experience working with various session properties to extract data from Salesforce object using standard api,
Bulk api
• Used IDQ’s standardized plans for addresses and names clean ups and complete initial data profiling and
removing duplicate data
• Addressed many performance issues on ETL jobs, semantic views, stored procedures, Reporting and Ad-hoc
SQL
• Worked in the Performance Tuning of SQL, ETL and other processes to optimize session performance
• Created Reusable transformations, Mapplets, Worklets using Transformation Developer, Mapplet Designer
and Worklet Designer
• Loaded data from Unstructured file format using unstructured data transformation into Oracle database
• Tuned Informatica mappings to improve the execution time by applying suitable Partitioning mechanisms and
tuning individual transformations inside the mapping
Sr. Data Specialist/ETL Developer - Lead
Project: Operational Data Storage and Data Warehousing Implementation
Client: Tufts Health Plan, Watertown, MA
Nov 2017 – Jan 2022
Tufts Health Plan is nationally recognized for its commitment to providing innovative, high-quality health care
coverage. Tufts Health Plan is one of the few health plans in Massachusetts to participate in the commercial, Medicare
and Medicaid/subsidized markets, offering coverage across the life span regardless of age or circumstance. ODS – Is
the Tufts Data Storage, the goal of Tufts Operational Data Storage (ODS) is to provide a centralized, high performance
repository that meets information needs across the organization. When paired with an Enterprise Business
Intelligence suite, the ODS provides a consistent version of the truth to the business, so reports and analytics can be
consistent, and results trusted.
Responsibilities:
• Interacting with the Business users to identify the process metrics and various key dimensions and measures
and involved in full life cycle of the project.
• Leading an On-Site and offshore team to Develop, Test, and Deploy the ETLs to Production in timely manner
• Interacting with SCRUM Masters and Agile Team to identify user stories and delivering the results within the
provided sprint time
• Assisting architect on Developing ODS / Dimensional warehouse logical and physical data models.
• Designed and developed ETL workflows using Informatica PowerCenter for data integration.
• Performed data profiling using Informatica Cloud Data Quality (CDQ) to identify data issues.
• Created and maintained technical documentation for ETL workflows, data models, and data mappings.
• Developing IDQ Rest Web Service API calls to external vendors to GET, PUT, and POST the Enterprise data by
converting them to JSON/XML format
• Using Salesforce Connectors to load Sales and Broker data from ODS to CRM reporting systems
• Using Address Validator in Data Quality to validate all the Enterprise wide member data
• Deriving required Fact and Dimension tables to accommodate the data as per to bring additional data
elements from Legacy RX source system into ODS as per the user requirements
• Developing Conceptual, Logical and Physical database models for ODS Phase-2 assurance applications and
forward engineered the DDLs
• Worked on Health rules Payor (HRP) and major clinical systems such as lab, pharmacy, radiology and EMR
applications
• Identifying ETL specifications based on Business Requirements and creating ETL Mapping Documents.
• Identifying issues in Production and fixing it on the Go
• Defined Modelling & Naming standards and Best Practices for the Modelling team to use in the Data models as
well as in the DDLs & DMLs while creating new data elements and adding attributes in ERStudio
16.0/17.0/19.0
• Identified and eliminated duplicates in datasets thorough IDQ 10.5 components of Edit Distance, Jaro Distance
and Mixed Field matcher, It enables the creation of a single view of customers, help control costs associated
with mailing lists by preventing multiple pieces of mail.
• Designed and developed the ETL processes using Informatica to load data from Oracle, SQL Server, Flat Files,
XML Files and JSON files to target Oracle Data Warehouse database
• Working closely with ETL team members to resolve the performance issues of ETL loads and business-related
questions.
Environment: Oracle 19c, SQL Server 2019, Elastic Search, ERStudio19, Informatica Power Center 10.5, IDQ 10.5, B2B,
Toad 11.6.0.43, SVN1.14
EIS is a contract being awarded by GSA for all US Federal Government Agencies. This is the replacement contract
vehicle for existing Networks, all GSA regional contracts such as Local Service Agreements (LSAs).
EIS transition refers to replacement service orders in EIS or transitioning services from existing contracts to EIS without
disruption of services. Verizon is supporting the EIS contract by using a hybrid architecture using proven VRD
(Verizon’s Rapid delivery and legacy infrastructure.
Responsibilities:
• Participating in requirement gathering meetings to identify functional, non-functional and performance
requirements of the project
• Gathering data from 1000+ legacy order / inventory / customer / billing and provisioning systems
• Assembling the collected data into a common representation of the legacy data
• Proposing ETL architecture solutions based on the requirements and working on High-level, Low-level Design
docs
• Responsible for designing and implementing strategic ETL processes using Informatica and PL/SQL framework.
• Designing and developing ETL mappings to load data from source to Target using Informatica Power Center
9.6.1/10.1.1
• Transforming the legacy data into VRD products and generating XREFs to compare legacy to VRD
• Loading the transformed data into data mart tables in GMART
• Involved in administration tasks including upgrading Informatica, importing/exporting mappings.
• Created and managed the multiple environments, user groups, backup and restore of the repositories
• Creating data manipulation and definition scripts using Teradata Bteq Utility.
• Using BTEQ scripts to load data from staging tables into Target tables.
• Working on MLOAD, FLOAD, TPUMP utilities of Teradata
• Loading unstructured data Via Informatica PC.
• Creating tables to load large sets of structured, semi-structured and unstructured data coming from UNIX,
NoSQL and a variety of portfolios.
• Assisting with data capacity planning and node forecasting
Environment: Informatica Power Center 9.6.1/10.1.1, Power Exchange, Oracle 11g, Teradata 15, Teradata Utilities,
Teradata SQL Assistant 15.10, Unix, WinScp 5.7.5, JIRA, BMC Control-M
ETL/Informatica SME(Remote)
Project: DDM- (Dynamic Data Masking)
Client: Freddie Mac, Reston, VA
Jan 2015 – Dec 2016
Freddie Mac, The Federal Home Loan Mortgage Corporation (FHLMC), is a public government-sponsored enterprise.
Freddie Mac has many applications, systems and tools that contain Protected Personal Information (collectively, “PPI
Applications.”) associated with over 60 million borrowers as well as for Freddie Mac employees, contingent workers
and other individuals. To address this security challenge, the PPI Applications – Data Protection Strategy project team
started implementing DDM project to protect high risk PPI data from unauthorized user access.
Informatica Dynamic Data Masking identifies data and controls unauthorized access to production environments. The
software dynamically masks sensitive information and blocks, audits, and alerts end users, IT personnel, and
outsourced teams who access sensitive information while ensuring compliance with privacy regulations.
Responsibilities:
• Participating in requirement gathering meetings to identify functional, non-functional and performance
requirements of the project
• Proposing DATA and ETL architecture solutions based on the requirements
• Translating requirements into solution architecture and design specifications that can be used for
development
• Responsible for designing and implementing strategic ETL processes with data privacy / data masking
solution(s)
• Identifying ETL specification based on Business Requirements/Mapping Document. Formulating and
documenting ETL process design. Participated and lead research and development efforts (proofs of concept,
prototypes), as subject matter experts, with new technologies.
• Designing and developing ETL mappings to load, transform data from source to Target using Informatica
Power Center 9.5.1
• Actively contributing and exploring data security / data privacy (data masking) technologies within the firm as
well as vendors and recommending appropriate solution
• Configuring DDM Connection rules, Security rules and AD Masking groups by views.
• Conducting daily status calls to gather status of the project and updating the project manager in a timely
manner
• Successfully implementing Dynamic Data Masking for various applications in different phases
Environment: Informatica Power Center 9.5.1, Teradata 14.0, Oracle 11g, SQL Server 2008, Sybase 15.7 Toad
11.6.0.43, Rapid SQL 8.1.1, Remedy, Service Now, Linux 6.7
ETL Developer
Project: NDW (National Data Warehouse)
Client: Lockheed Martin, Baltimore MD
Oct 2014 – Dec 2014
Centers for Medicare and Medicaid Services (CMS)
CMS is the Centers for Medicare & Medicaid Services and the federal agency responsible for administering the
Medicare, Medicaid, CHIP (Children's Health Insurance), HIPAA and several other health-related programs. The
purpose of the NDW responsible for providing user friendly access to CMS 1-800-MEDICARE information. NDW users
will be able to leverage a flexible and user-friendly web-based environment to produce reports for conducting data
analysis. NDW collect all data
from CMS help line operational system into a single data warehouse and then further organize it into small data marts
based on the different subject areas as per business requirement.
Responsibilities:
• Interacted with business analysts and developers to analyze the user requirements, functional specifications,
and system specifications.
• Used Data Quality 8.6.1 (IDQ) toolkit, Analysis, data cleansing, data matching, data conversion, exception
handling, and reporting and monitoring capabilities of IDQ 8.6.1.
• Proficiently worked on Conceptual, Physical and Logical data models (3NF) using various Data Modeling tools
like Erwin and MS Visio with strong understanding in the principles of Data ware housing using Fact Tables,
Dimension Tables, star schema modeling and snowflake schema modeling, foreign key concepts, referential
integrity.
• Identified and tracked the slowly changing dimensions/mini dimensions, heterogeneous Sources and
determined the hierarchies in dimensions.
• Created application-specific Data Marts so that users can access personalized dashboards of information that
is specific to their department and business unit.
• Worked on Query optimization & Performance tuning using Execution Plan, Performance Monitor.
• Identified ETL specification based on Business Requirements/Mapping Document and formulate process
design.
• Implemented ETL standards and Best Practices while naming Maps and processes.
• Worked closely with all team members to resolve the technical issues and business-related questions.
Environment: Oracle 11g, Flat Files, Teradata, Erwin r7.3.3, Informatica Power Center 8.6.1, Data Profiling / IDQ8.6.1,
Toad 9.7.2, SQL, PL/SQL, WIN SCP 4.3.4, Unix Shell Scripts, Quality Center 9.0, Sun Solaris 10.
CERTIFICATIONS
• Informatica Certified Professional
EDUCATION
• Master of Computer Science - Year 2017, University of Springfield, Illinois, USA
• Bachelor of Science in Engineering - Year 2012, Osmania University, India