Data Modeling

Download as pptx, pdf, or txt
Download as pptx, pdf, or txt
You are on page 1of 41

MODULE 7:

Data Modeling
LEADER:
GONATISE, CARL

MEMBERS:
GONGORA, JORIZ
GUIUO, MARLEN
JAMIAS, MARLYN
INTRODUCTION TO DATA
MODELING
• Data modeling is the process of creating a conceptual
representation of real-world entities and their relationships
within a particular domain. It is crucial for database design,
software development, and data management. Data models
visually represent information systems, showing how data is
used, stored, organized, and related. These models are built
based on business needs and stakeholder feedback, and
evolve with changing business requirements.
INTRODUCTION TO DATA
MODELING
• Data modeling uses standardized schemas and techniques to
maintain consistency. There are different types of models:
conceptual models offer a high-level overview, while logical
and physical models focus on implementation. Techniques
like Entity-Relationship Modeling (ER) and Unified Modeling
Language (UML) help graphically represent entities,
attributes, and relationships.
What Makes a
Good Model?
What Makes a Good Model?
1. Keep Business Objectives in Mind:
 Always start with a clear understanding of the business
requirements and objectives.
 Consider how the data model will support business
processes, reporting, and analytics.
 Align your data model design with the organization’s
strategic goals
What Makes a Good Model?
2. Properly Document Your Data Model:
 Comprehensive documentation is essential for
understanding and maintaining data models.
 Document the purpose of each table, relationships,
attributes, and any assumptions made during modeling.
 Clear documentation aids collaboration among
stakeholders.
What Makes a Good Model?
3. Design Your Data Model to Be Adjustable Over Time:
 Business needs evolve, and data models should be flexible
enough to accommodate changes.
 Avoid rigid structures that hinder adaptability.
 Plan for scalability and future enhancements.
What Makes a Good Model?
4. Have Different Levels of Abstraction for Different People:
 Data models serve various audiences: business users,
analysts, developers, and database administrators.
 Provide high-level conceptual models for executives and
detailed logical/physical models for technical teams.
 Tailor the level of detail based on the audience’s expertise.
What Makes a Good Model?
5. Choose the Right Data Modeling Technique:
 Use standard notations such as Entity-Relationship
Diagrams (ERDs) or Unified Modeling Language (UML).
 ERDs visually represent entities, attributes, and
relationships.
 UML diagrams provide a broader view, including use cases,
classes, and interactions.
What Makes a Good Model?
6. Think about Data Governance and Security:
 Incorporate data governance principles into your
modeling process.
 Define access controls, data ownership, and privacy rules.
 Ensure compliance with regulations (e.g., GDPR, HIPAA).
What Makes a Good Model?
7. Avoid Premature Optimization:
 Focus on meeting business requirements first.
 Optimize for performance only when necessary.
 Premature optimization can lead to complex and over-
engineered models.
TYPES OF DATA MODELS
1. Conceptual Data Model
A conceptual data model (CDM) is a high-level
representation of the business requirements and the
connected data sets and relationships, independent of any
specific implementation or technology
The Entity-Relationship Diagram (ERD) is a graphical
representation of the conceptual data model, depicting entities,
attributes, and relationships:
2. Logical Design Stage
A logical data model (LDM) is an abstract representation of a possible
implementation, without being bound to any specific implementation or
technology.

The logical design stage translates the conceptual model into a logical
data model, following specific data modeling rules and conventions:
3. Physical Data Model
A physical data model (PDM) is a detailed representation of how the data
is actually stored and manipulated in a specific system.

In a physical data model, the entities are transformed into tables, while
the attributes are transformed into columns:
DATA MODELING TECHNIQUES
Enterprise Data Modeling

Entity-Relationship
Modeling (ER Modeling) is
a widely used data
modeling technique,
especially for relational
database design. It
represents real-world
objects or concepts as
entities and their
associations as
relationships. ER models
use graphical notations,
such as rectangles for
entities, diamonds for
relationships, and
connecting lines.
Unified Modeling Language
(UML)
Unified Modeling Language
(UML) is a standardized
modeling language used in
software engineering and
system design. It includes
various diagram types, with
class diagrams being useful
for data modeling. Class
diagrams represent entities
(classes) along with their
attributes, operations, and
relationships. UML also
supports advanced concepts
like inheritance,
composition, and
aggregation, offering a
comprehensive and flexible
approach to modeling both
data and behavior.
Data Flow Diagrams (DFDs)
Data Flow Diagrams (DFDs)
are used in structured
systems analysis to model
the flow of data through a
system and the processes
that transform it. They
consist of external entities,
data stores, processes, and
data flows, helping to clarify
data requirements and
system interactions.
Object-Role Modeling (ORM)
Object-Role Modeling (ORM)
is a fact-based data
modeling approach that
represents information using
natural language sentences.
It captures business rules
and constraints through
predicate logic and set
theory, supporting advanced
concepts like subtypes,
mandatory roles, and value
constraints. ORM is
particularly useful for
capturing complex business
rules and ensuring data
integrity.
Dimension
Modeling
Dimension modeling is a
technique designed for data
warehousing and business
intelligence. It organizes
data into facts (measurable
metrics) and dimensions
(descriptive attributes) and
uses star or snowflake
schemas for efficient
querying. This approach
focuses on supporting
analytical queries and
reporting while addressing
the integration of data from
multiple sources.
DATA MODELING PROCESS
DATA MODELING PROCESS
Data modeling techniques have
different conventions that dictate
which symbols are used to represent
the data, how models are laid out, and
how business requirements are
conveyed. All approaches provide
formalized
workflows that include a sequence
of tasks to be performed in an iterative
manner.
DATA MODELING PROCESS
• Identify the entities. The process of data modeling begins with the
identification of the things, events or concepts that are represented in the
data set that is to be modeled. Each entity should be cohesive and logically
discrete from all others.

• Identify key properties of each entity. Each entity type can be


differentiated from all others because it has one or more unique properties,
called attributes. For instance, an entity called “customer” might possess
such attributes as a first name, last name, telephone number and salutation,
while an entity called “address” might include a street name and number, a
city, state, country and zip code.
DATA MODELING PROCESS
• Identify relationships among entities. The earliest draft of a data
model will
specify the nature of the relationships each entity has with the others. In the
above example, each customer “lives at” an address. If that model were
expanded to include an entity called “orders,” each order would be shipped
to and billed to an address as well. These relationships are usually
documented via unified modeling language (UML).

• Map attributes to entities completely. This will ensure the model


reflects how the business will use the data. Several formal data modeling
patterns are in widespread use. Object-oriented developers often apply
analysis patterns or design patterns, while stakeholders from other business
domains may turn to other patterns.
DATA MODELING PROCESS
• Assign keys as needed, and decide on a degree of normalization
that
balances the need to reduce redundancy with performance
requirements.
Normalization is a technique for organizing data models (and the databases
they represent) in which numerical identifiers, called keys, are assigned to
groups of data to represent relationships between them without repeating the
data.

• Finalize and validate the data model. Data modeling is an iterative


process that should be repeated and refined as business needs change.
ENTERPRISE DATA MODELING
ENTERPRISE DATA MODELING
Enterprise data modeling involves
developing a comprehensive data model
representing an organization's entire data
and its relationships. It addresses data
integration and sharing challenges across
different systems and departments. Data
governance and quality management are
crucial to maintaining the integrity and
consistency of these models. The
enterprise data model must align with the
organization's business goals, strategies,
and data management policies.
PROBLEM OF DATA MANAGEMENT
PROBLEM OF DATA MANAGEMENT
Problem of Data Management:

• Managing data as a shared resource across an organization requires coordination,


governance, and consistent practices.

• Data redundancy and inconsistencies can arise when data is duplicated or maintained
separately in different systems or departments.

• Data redundancy and inconsistencies can arise when data is duplicated or maintained
separately in different systems or departments.

• Data silos and lack of integration can hinder data sharing, analysis, and decisionmaking
processes.

• Data redundancy and inconsistencies can arise when data is duplicated or maintained
separately in different systems or departments.

• Data silos and lack of integration can hinder data sharing, analysis, and decisionmaking
processes.
PROBLEM OF DATA MANAGEMENT
Problem of Data Management:

• Data quality issues, such as incomplete, inaccurate, or outdated data, can have
significant impacts on business operations and decision-making.

• Data redundancy and inconsistencies can arise when data is duplicated or maintained
separately in different systems or departments.

• Data silos and lack of integration can hinder data sharing, analysis, and decisionmaking
processes.

• Data quality issues, such as incomplete, inaccurate, or outdated data, can have
significant impacts on business operations and decision-making.

• Data cleansing and data quality management processes are essential for ensuring the
reliability and usability of data.
MANAGING DATA AS SHARED
RESOURCE
MANAGING DATA AS SHARED RESOURCE
Managing data as shared data refers to the practice of treating data as a valuable, common
resource that can be accessed, used, and maintained across different systems, departments,
and stakeholders within an organization.

This approach promotes data integration, consistency, and collaboration. Key aspects of
managing data as shared data include:

 Data Governance: Establishing rules and policies for managing data


access, quality, security, and compliance. This ensures that shared data
is accurate, trustworthy, and used responsibly.

 Data Integration: Ensuring that data from different sources


(departments, systems, or external entities) is harmonized and
integrated, so it can be used across the organization without silos.

 Data Accessibility: Making sure that shared data is available to


authorized users in a usable format. This may involve central
repositories or data platforms where multiple users can access the same
MANAGING DATA AS SHARED RESOURCE
 Data Quality: Monitoring and maintaining the quality of shared data by
ensuring consistency, accuracy, and completeness. High-quality data
supports better decision-making.

 Collaboration and Transparency: Encouraging different departments


or teams to share insights and use the same data, promoting
transparency and reducing redundant efforts.

 Security and Privacy: Implementing strong security measures to


protect sensitive or confidential data while enabling sharing in a
controlled and secure manner.

 Data Stewardship: Designating data stewards or managers responsible


for ensuring that data is well-maintained, accurate, and available for all
relevant users.
DEVELOPMENT CYCLE
DEVELOPMENT CYCLE
The development cycle refers to the structured process or series of stages that
a project or product goes through from initial concept to completion. It outlines
the phases of development, ensuring that tasks are completed in an orderly,
efficient, and systematic manner.

Key Phases of the Development


Cycle:

• Planning/Requirements Gathering
• Analysis
• Design
• Development/Implementation
• Testing/Quality Assurance
• Implementation/Release
• Maintenance/Support
EXPERTISE REQUIREMENTS
Here are some key expertise areas:

 SQL Proficiency: Being able to write and optimize SQL queries to interact with databases

 Database Design: Knowledge of designing efficient and scalable database structures

 Data Warehousing: Familiarity with data warehousing concepts and technologies

 ETL Processes: Experience with Extract, Transform, Load (ETL) processes to move data
between systems

 Dimensional Modeling: Skills in designing dimensional models for data warehousing

 Metadata Management: Managing metadata to ensure data consistency and integrity


EXTERNAL STANDARDS
 TOGAF (The Open Group Architecture Framework): This framework provides a detailed
method and a set of supporting tools for developing an enterprise architecture, including data
models.

 Data Vault Modeling: This methodology is used for designing enterprise data warehouses
and focuses on long-term historical storage of data.

 Star Schema and Snowflake Schema: These are common dimensional modeling techniques
used in data warehousing to organize data into fact and dimension tables.
THANKYOU FOR
LISTENING

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy