Cse III Discrete Mathematical Structures 10cs34 Notes
Cse III Discrete Mathematical Structures 10cs34 Notes
UNIT – 1 6 Hours
Set Theory: Sets and Subsets, Set Operations and the Laws of Set Theory, Counting and
Venn Diagrams, A First Word on Probability, Countable and Uncountable Sets
UNIT – 2 7 Hours
Fundamentals of Logic: Basic Connectives and Truth Tables, Logic Equivalence – The
Laws of Logic, Logical Implication – Rules of Inference
UNIT – 3 6 Hours
Fundamentals of Logic contd.: The Use of Quantifiers, Quantifiers, Definitions and the
Proofs of Theorems
UNIT – 4 7 Hours
Properties of the Integers: Mathematical Induction, The Well Ordering Principle –
Mathematical Induction, Recursive Definitions
PART – B
UNIT – 5 7 Hours
Relations and Functions: Cartesian Products and Relations, Functions – Plain and One-to-
One, Onto Functions – Stirling Numbers of the Second Kind, Special Functions, The
Pigeon-hole Principle, Function Composition and Inverse Functions
UNIT – 6 7 Hours
Relations contd.: Properties of Relations, Computer Recognition – Zero-One Matrices and
Directed Graphs, Partial Orders – Hasse Diagrams, Equivalence Relations and Partitions
UNIT – 7 6 Hours
Groups: Definitions, Examples, and Elementary Properties, Homomorphisms,
Isomorphisms, and Cyclic Groups, Cosets, and Lagrange‘s Theorem.
Coding Theory and Rings: Elements of Coding Theory, The Hamming Metric, The Parity
Check, and Generator Matrices
UNIT – 8 6 Hours
Group Codes: Decoding with Coset Leaders, Hamming Matrices Rings and Modular
Arithmetic: The Ring Structure – Definition and Examples, Ring Properties and
Substructures, The Integers Modulo n
Text Book:
1. Ralph P. Grimaldi: Discrete and Combinatorial Mathematics,5 Edition, Pearson
Education, 2004. (Chapter 3.1, 3.2, 3.3, 3.4, Appendix 3, Chapter 2, Chapter 4.1, 4.2,
Chapter 5.1 to 5.6, Chapter 7.1 to 7.4, Chapter 16.1, 16.2, 16.3, 16.5 to 16.9, and Chapter
14.1, 14.2, 14.3).
Reference Books:
1. Kenneth H. Rosen: Discrete Mathematics and its Applications, 7 Edition, McGraw Hill,
2010.
2. Jayant Ganguly: A Treatise on Discrete Mathematical Structures, Sanguine-Pearson,
2010.
3. D.S. Malik and M.K. Sen: Discrete Mathematical Structures: Theory and Applications,
Cengage Learning, 2004.
4. Thomas Koshy: Discrete Mathematics with Applications, Elsevier, 2005, Reprint 2008.
INDEX SHEET
Groups: 75 - 90
UNIT – 7
Coding Theory and Rings: 91 - 103
PART – A
DISRETE MATHEMATICAL
STRUCTURES
UNIT I 6 Hours
Set Theory
Sets: A set is a collection of objects, called elements of the set. A set can be
presented by listing its elements between braces: A = {1, 2, 3, 4, 5}. The symbol e is
used to express that an element is (or belongs to) a set. For instance 3 e A. Its
negation is represented by /e, e.g. 7 /e A. If the set Is finite, its number of elements
is represented |A|, e.g. if A = {1, 2, 3, 4, 5} then |A| = 5
Set-builder notation: An alternative way to define a set, called set- builder notation, is
by stating a property (predicate) P (x) verified by exactly its elements, for instance
A = {x e Z | 1 ≤ x ≤ 5} = ―set of integers x such that 1 ≤ x ≤ 5‖—i.e.: A = {1, 2, 3,
4, 5}. In general: A = {x e U | p(x)}, where U is the universe of discourse in which
the predicate P (x) must be interpreted, or A = {x | P (x)} if the universe of discourse
for P (x) is implicitly understood. In set theory the term universal set is often used in
place of ―universe of discourse‖ for a given predicate.
Principle of Extension: Two sets are equal if and only if they have the same
elements, i.e.: A = B ≡ ∀x (x e A ↔ x e B) .
Empty Set: A set with no elements is called empty set (or null set,
or void set ), and is represented by ∅ or {}.
Note that nothing prevents a set from possibly being an element of another set (which
is not the same as being a subset!). For i n stance
if A = {1, a, {3, t}, {1, 2, 3}} and B= {3, t}, then obviously B is an element of A,
i.e., B e A.
Power Set: The collection of all subsets of a set A is called the power set of A,
and is represented P(A). For instance, if A = {1, 2, 3}, then
P(A) = {∅, {1}, {2}, {3}, {1, 2}, {1, 3}, {2, 3}, A} .
Multisets: Two ordinary sets are identical if they have the same elements, so for
instance, {a, a, b} and {a, b} are the same set because they have exactly the same
elements, namely a and b. However, in some applications it might be useful to
allow repeated elements in a set. In that case we use multisets, which are
mathematical entities similar to sets, but with possibly repeated elements. So, as
multisets, {a, a, b} and {a, b} would be considered different, since in the first one the
element a occurs twice and in the second one it occurs only once.
Set Operations:
3. Complement : The set of elements (in the universal set) that do not belong to a
given set:
A = {x e U | x /e A} .
4. Difference or Relative Complement : The set of elements that belong to a set but
not to another:
A − B = {x | (x e A) ∧ (x /e B)} = A ∩ B .
5. Symmetric Difference : Given two sets, their symmetric differ- ence is the set of
elements that belong to either one or the other set but not both.
A ⊕ B = {x | (x e A) ⊕ (x e B)} .
Example : Let M , P and C be the sets of students taking Mathe- matics courses,
Physics courses and Computer Science courses respec- tively in a university. Assume
|M | = 300, |P | = 350, |C | = 450,
|M ∩ P | = 100, |M ∩ C | = 150, |P ∩ C | = 75, |M ∩ P ∩ C | = 10. How
many students are taking exactly one of those courses? (fig. 2.7)
We see that |(M ∩P )−(M ∩P ∩C )| = 100−10 = 90, |(M ∩C )−(M ∩
P ∩ C )| = 150 − 10 = 140 and |(P ∩ C ) − (M ∩ P ∩ C )| = 75 − 10 = 65.
Then the region corresponding to students taking Mathematics courses only has
cardinality 300−(90+10+140) = 60. Analogously we compute the number of students
taking Physics courses only (185) and taking Computer Science courses only (235).
The sum 60 + 185 + 235 = 480 is the number of students taking exactly one of those
courses.
Venn Diagrams:
Venn diagrams are graphic representa- tions of sets as enclosed areas in the plane.
For instance, in figure 2.1, the rectangle represents the universal set (the set of all
elements con- sidered in a given problem) and the shaded region represents a set A.
The other figures represent various set operations.
Analogously, their intersection is the set of elements that belong to all the sets
simultaneously:
These definitions can be applied to infinite collections of sets as well. For instance
assume that Sm = {kn | k = 2, 3, 4, . . . } = set of multiples of n greater than n. Then
Example : The division of the integers Z into even and odd numbers is a partition: S =
{E, O}, where E = {2n | n e Z}, O = {2n + 1 | n e Z}.
If all possible outcomes of an experiment have the same likelihood of occurrence, then
the probability of an event A ⊂ S is given by Laplace‘s rule:
ADDITION THEROM:
Suppose A and B are 2 events is a sample space S then A UB is an event in S consisting of
outcomes that are in A or B or both and A ∩ B is an event is S consisting of outcomes
thatarecommon to A and B. accordingly by the principle of addition we have
|AUB|=|A|+|B|-|A ∩B| and formula 1 gives
Pr(AUB)=|AUB|/|S|=(|A|+|B|-|A ∩B|)/|S|
= |A|/|S| + |B|/|S| - |A ∩ B| / |S|
Pr(AUB) =Pr(A)+Pr(B)-Pr(A ∩ B)
CONDITIONAL PROBABILITY:
If E is an event in a finite sample S with Pr(E)>0 then the probability that an event A in S
occurs when E has already occurred is called the probability of A relative to E or the
conditional probability of A , given E
This probability, denoted by Pr(A|E) is defined by
Pr(A|E)=|A∩ E|/ |E| from this |A∩ E|=|E| . Pr(A|E)
Pr(A∩ E)= |A∩ E|/ S=|=|E|/|S| . Pr(A|E)=Pr(E) . Pr(A|E)
Example : Find the probability of obtaining a sum of 10 after rolling two fair dice. Find
the probability of that event if we know that at least one of the dice shows 5 points.
Answer : We call E — ―obtaining sum 10‖ and F — ―at least one of the dice shows 5
points‖. The number of possible outcomes is 6 × 6 —
36. The event ―obtaining a sum 10‖ is E — {(4, 6), (5, 5), (6, 4)}, so
|E| — 3. Hence the probability is P (E) — |E|/|S| — 3/36 — 1/12.
Now, if we know that at least one of the dice shows 5 points then the
sample space shrinks to
F — {(1, 5), (2, 5), (3, 5), (4, 5), (5, 5), (6, 5), (5, 1), (5, 2), (5, 3), (5, 4), (5, 6)} ,
so |F | — 11, and the ways to obtain a sum 10 are E n F — {(5, 5)},
|E n F | — 1, so the probability is P (E | F ) — P (E n F )/P (F ) — 1/11.
1. The probability that the shooter does not hit the target in one shot.
2. The probability that the shooter does not hit the target three times in a row.
3. The probability that the shooter hits the target at least once after shooting three
times.
Answer :
0.973.
UNIT 2 7 Hours
Fundamentals of Logic
SYLLABUS
Fundamentals of Logic: Basic Connectives and Truth Tables, Logic
Equivalence – The Laws of Logic, Logical Implication – Rules of
Inference
Introduction:
Propositions:
A proposition is a declarative sentence that is either true or false (but not both). For
instance, the following are propositions: ―Paris is in France‖ (true), ―London is in
Denmark‖ (false), ―2 < 4‖ (true), ―4 = 7 (false)‖. However the following are not
propositions: ―what is your name?‖ (this is a question), ―do your homework‖ (this
is a command), ―this sentence is false‖ (neither true nor false), ―x is an even
number‖ (it depends on what x represents), ―Socrates‖ (it is not even a sentence).
The truth or falsehood of a proposition is called its truth value.
Basic Connectives and Truth Tables:
Connectives are used for making compound propositions. The main ones are
the following (p and q represent given propositions):
The truth value of a compound proposition depends only on the value of its
components. Writing F for ―false‖ and T for ―true‖, we can summarize the meaning
of the connectives in the following way:
Note that ∨ represents a non-exclusive or, i.e., p ∨ q is true when any of p, q is true
and also when both are true. On the other hand ⊕ represents an exclusive or, i.e., p
⊕ q is true only when exactly one of p and q is true.
p ¬p p ∨ ¬p p ∧ ¬p
T F T F
T F T F
F T T F
F T T F
✻ ✻
tautology contradiction
Note that p → q is true always except when p is true and q is false. So, the following
sentences are true: ―if 2 < 4 then Paris is in France‖ (true → true), ―if London is
in Denmark then 2 < 4‖ (false → true),
―if 4 = 7 then London is in Denmark‖ (false → false). However the following one
is false: ―if 2 < 4 then London is in Denmark‖ (true → false).
of the truth value of q. This will become clearer when we study predicates such as ―if
x is a multiple of 4 then x is a multiple of 2‖. That implication is obviously true,
although for the particular
case x = 3 it becomes ―if 3 is a multiple of 4 then 3 is a multiple of 2‖.
The proposition p ↔ q, read ―p if and only if q‖, is called bicon- ditional. It is true
precisely when p and q have the same truth value, i.e., they are both true or both
false.
p q ¬p ¬p ∨ q p →
T T F T T
q
T F F F F
F T T T T
F F T T T
When two compound propositions have the same truth values no matter what truth
value their constituent propositions have, they are called logically equivalent. For
instance p → q and ¬p ∨ q are logically equivalent, and we write it:
p → q ≡ ¬p ∨ q
Note that that two propositions A and B are logically equivalent precisely when A ↔
B is a tautology.
Example : De Morgan‘s Laws for Logic. The following propositions are logically
equivalent:
¬(p ∨ q) ≡ ¬p ∧ ¬q
¬(p ∧ q) ≡ ¬p ∨ ¬q
p ↔ q ≡ (p → q) ∧ (q → p)
p q p → q → (p → q) ∧ (q p ↔ q
T T T
q T
p T T
→ p)
T F F T F F
F T T F F F
F F T T T T
¬(p → q) ≡ p ∧ ¬q
p → q ≡ ¬q → ¬p
¬(p ↔ q) ≡ p ⊕ q
p ↔ q ≡ (p → q) ∧ (q → p)
So, for instance, saying that ―John is married if and only if he has a spouse‖ is the
same as saying ―if John is married then he has a spouse‖ and ―if he has a spouse then
he is married‖.
Note that the converse is not equivalent to the given conditional proposition, for
instance ―if John is from Chicago then John is from Illinois‖ is true, but the converse
―if John is from Illinois then John is from Chicago‖ may be false.
The contrapositive of a conditional proposition p → q is the propo- sition ¬q →
¬p. They are logically equivalent. For instance the con- trapositive of ―if John is
from Chicago then John is from Illinois‖ is ―if
John is not from Illinois then John is not from Chicago‖.
LOGICAL CONNECTIVES: New propositions are obtained with the aid of
word or phrases like ―not‖,‖and‖,‖if….then‖,and ―if and only if‖. Such words or phrases
are called logical connectives. The new propositions obtained by the use of connectives are
Dept. of CSE, SJBIT Page 20
DMS 10CS34
Truth Table: p ~p
0 1
1 0
CONJUNCTION:
A compound proposition obtained by combining two given propositions by inserting the
word ―and‖ in between them is called the conjunction of the given proposition.The
conjunction of two proposition p and q is denoted by p^q(read ―p and q‖).
• The conjunction p^q is true only when p is true and q is true; in all other cases it is
false.
• Ex: p:2 is an irational number q: 9 is a prime number
p^q: 2 is an irational number and 9 is a prime number
• Truth table: p q p^q
0 0 0
0 1 0
1 0 0
1 1 1
DISJUNCTION:
A compound proposition obtained by combining two given propositions by inserting the
word ―or‖ in between them is called the disjunction of the given proposition.The
disjunction of two proposition p and q is denoted by pq(read ―p or q‖).
• The disjunction pq is false only when p is false and q is false ; in all other cases it
is true.
• p q pq
0 0 0
0 1 1
1 0 1
1 1 1
EXCLUSIVE DISJUNCTION:
• The compound proposition ―p or q‖ to be true only when either p is true or q is true
but not both. The exclusive or is denoted by symbol v.
• Truth Table:
p q pvq
0 0 0
0 1 1
1 0 1
1 1 0
CONDITIONAL(or IMPLICATION):
• A compound proposition obtained by combining two given propositions by using
the words ―if‖ and ―then‖ at appropriate places is called a conditional or an implication..
• Given two propositions p and q, we can form the conditionals ―if p, then q‖ and ―if
q, then p:. The conditional ―if p, then q‖is denoted by pq and the conditional ―if q, then
p‖ is denoted by q p.
• The conditional pq is false only when p is true and q is false ;in all other cases it
is true.
• Truth Table:
p q pq
0 0 1
0 1 1
1 0 0
1 1 1
BICONDITIONAL:
• Let p and q be two propositions,then the conjunction of the conditionals pq and
qp is called bi-conditional of p and q. It is denoted by pq.
A compound proposition which is always false regardless of the truth values of its
components is called a contradiction or an absurdity.
A compound proposition that can be true or false (depending upon the truth values of its
components) is called a contingency I.e contingency is a compound proposition which is
neither a tautology nor a contradiction.
LOGICAL EQUIVALENCE
• Two propositions ‗u‘ and ‗v‘ are said to be logically equivalent whenever u and v
have the same truth value, or equivalently .
• Then we write uv. Here the symbol stands for ―logically equivalent to‖.
• When the propositions u and v are not logically equivalent we write uv.
LAWS OF LOGIC:
To denote a tautology and To denotes a contradiction.
• Absorption Laws: For any proposition p and q,1) [p (pq)] p 2)[p (pq)] p
• Law for the negation of a conditional : Given a conditional pq, its negation is
obtained by using the following law: (pq)[p(q)]
NOTE:
• ~ (pq) pq
• (pq) p q
• (pq) [p(q)]
• Ex: current flows from the terminal A to the terminal B if the switch is closed i.e if
p is assigned the symbol 1. This network is represented by the symbol p
A P B
• Ex: parallel network consists of 2 switches p and q in which the current flows from
the terminal A to the terminal B , if p or q or both are closed i.e if p or q (or both) are
assigned the symbol 1. This network is represent by pq
Ex: Series network consists of 2 switches p and q in which the current flows from
the terminal A to the terminal B if both of p and q are closed; that is if both p and q are
assigned the symbol 1. This network is represented by pq
DUALITY:
Suppose u is a compound proposition that contains the connectives and . Suppose we
replace each occurrence of and in u by and respectively.
NOTE:
• (ud)d u. The dual of the dual of u is logically equivalent to u.
• For any two propositions u and v if u v, then ud vd . This is known as the
principle of duality.
(p q)= (p q) p q
(p q)= (p q) p q
CONVERSE,INVERSE AND CONTRAPOSITIVE
Consider a conditional (pq) , Then :
RULES OF INFERENCE:
There exist rules of logic which can be employed for establishing the validity of arguments
. These rules are called the Rules of Inference.
1) Rule of conjunctive simplification: This rule states that for any two propositions p
and q if pq is true, then p is true i.e (pq )p.
2) Rule of Disjunctive amplification: This rule states that for any two proposition p
and q if p is true then pq is true i.e p(p q)
3) 3) Rule of Syllogism: This rule states that for any three propositions p,q r if pq is
true and qr is true then pr is true. i.e {(pq)(q)} (p r) In tabular form:
pq qr (p r)
4) 4) Modus pones(Rule of Detachment): This rule states that if p is true and pq
is true, then q is true, ie {p (pq )} q. Tabular form
p pq q
5) Modus Tollens: This rule states that if pq is true and q is false, then p is false.
{(pq)q}q Tabular form: pq
q p
6) Rule of Disjunctive Syllogism: This rule states that if pq is true and p is false,
then q is true i.e. {(pq)p}q Tabular Form pq
p q
QUANTIFIERS:
1. The words ―ALL‖,‖EVERY‖,‖SOME‖,‖THERE EXISTS‖ are called quantifiers in
the proposition
LOGICAL EQUIVALENCE:
1. x,[p(x)q(x)](x p(x))(x,q(x))
{x,p(x)}x{p(x)} {x,p(x)}x{p(x)}
RULES OF INTERFERENCE:
1. RULE OF UNIVERSAL SPECIFICATION
If an open statement p(x) is proved to be true for any (arbitrary)x chosen from a set S,
then the quantified statement x€S, p(x) is true.
b) Analysis: starting with the hypothesis and employing the rules /laws of
logic and other known facts , infer that q is true
c) Conclusion:pq is true.
2. INDIRECT PROOF:
Condition pq and its contrapositive qp are logically equivalent. On basis of this
proof, we infer that the conditional pq is true. This method of proving a conditional is
called an indirect method of proof.
3.PROOF BY CONTRADICTION
The indirect method of proof is equivalent to what is known as the proof by contradiction.
The lines of argument in this method of proof of the statement pq are as follows:
1) Hypothesis: Assume that pq is false i.e assume that p is true and q is
false.
2)Analysis: starting with the hypothesis that q is false and employing the rules of
logic and other known facts , infer that p is false. This contradicts the assumption that p is
true
4.PROOF BY EXHAUSTION:
―x €S,p(x)‖ is true if p(x)is true for every (each) x in S.If S consists of only a limited
number of elements , we can prove that the statement ―x €S,p(x)‖ is true by considering
p(a) for each a in S and verifying that p(a) is true .such a method of prove is called method
of exhaustion.
5.PROOF OF EXISTENCE:
―x €S,p(x)‖ is true if any one element a € S such that p(a) is true is exhibited. Hence , the
best way of proving a proposition of the form ―x €S,p(x)‖ is to exhibit the existence of
one a€S such that p(a) is true. This method of proof is called proof of existence.
6.DISPROOF BY CONTRADICTION :
Suppose we wish to disprove a conditional pq. for this propose we start with the
hypothesis that p is true and q is true, and end up with a contradiction. In view of the
contradiction , we conclude that the conditional pq is false.this method of disproving
pq is called DISPROOF BY CONTRADICTION
―x €S,p(x)‖ is false if any one element a€S such that p(a) is false is exhibited hence the
best way of disproving a proposition involving the universal quantifiers is to exhibit just
one case where the proposition is false. This method of disproof is called DISPROOF BY
COUNTER EXAMPLE
Unit 3 6 Hours
Fundamentals of Logic contd.:
SYLLABUS
Fundamentals of Logic contd.: The Use of Quantifiers, Quantifiers,
Definitions and the Proofs of Theorems
Predicates, Quantifiers
Predicates:
A predicate or propositional function is a state- ment containing variables. For instance
―x + 2 = 7‖, ―X is American‖, ―x < y‖, ―p is a prime number‖ are predicates. The
truth value of the predicate depends on the value assigned to its variables. For instance
if we replace x with 1 in the predicate ―x + 2 = 7‖ we obtain ―1 + 2 = 7‖, which is false,
but if we replace it with 5 we get ―5 + 2 = 7‖, which is true. We represent a
predicate by a letter followed by the variables enclosed between parenthesis: P (x),
Q(x, y), etc. An example for P (x) is a value of x for which P (x) is true. A
counterexample is a value of x for which P (x) is false. So, 5 is an example for ―x + 2
= 7‖, while 1 is a counterexample.
Quantifiers:
Given a predicate P (x), the statement ―for some x, P (x)‖ (or ―there is some x
such that p(x)‖), represented ―∃x P (x)‖, has a definite truth value, so it is a
proposition in the usual sense. For instance if P (x) is ―x + 2 = 7‖ with the integers
as
universe of discourse, then ∃x P (x) is true, since there is indeed an integer, namely
5, such that P (5) is a true statement. However, if
Q(x) is ―2x = 7‖ and the universe of discourse is still the integers, then ∃x Q(x) is
false. On the other hand, ∃x Q(x) would be true if we extend the universe of
discourse to the rational numbers. The symbol
Analogously, the sentence ―for all x, P (x)‖—also ―for any x, P (x)‖, ―for every x, P (x)‖,
―for each x, P (x)‖—, represented ―∀x P (x)‖, has a definite truth value. For instance,
if P (x) is ―x + 2 = 7‖ and the
universe of discourse is the integers, then ∀x P (x) is false. However if Q(x) represents
―(x + 1)2 = x2 + 2x + 1‖ then ∀x Q(x) is true. The symbol ∀ is called the universal
quantifier.
In predicates with more than one variable it is possible to use several quantifiers at the
same time, for instance ∀x∀y∃z P (x, y, z), meaning ―for all x and all y there is some z
such that P (x, y, z)‖.
Note that in general the existential and universal quantifiers cannot be swapped, i.e.,
in general ∀x∃y P (x, y) means something different from ∃y∀x P (x, y). For instance if
x and y represent human beings and P (x, y) represents ―x is a friend of y‖, then ∀x∃y
P (x, y) means that everybody is a friend of someone, but ∃y∀x P (x, y) means that
there is someone such that everybody is his or her friend.
A predicate can be partially quantified, e.g. ∀x∃y P (x, y, z, t). The variables quantified
(x and y in the example) are called bound variables, and the rest (z and t in the
example) are called free variables. A
partially quantified predicate is still a predicate, but depending on
fewer variables.
On the other hand, if ∀x P (x) is false then it is not true that for every x, P (x) holds,
hence for some x, P (x) must be false. Thus:
¬∀x P (x) ≡ ∃x ¬P (x) .
This two rules can be applied in successive steps to find the negation of a more complex
quantified statement, for instance:
¬∃x∀y p(x, y) ≡ ∀x¬∀y P (x, y) ≡ ∀x∃y ¬P (x, y) .
Exercise : Write formally the statement ―for every real number there is a greater real
number‖. Write the negation of that statement.
Answer : The statement is: ∀x ∃y (x < y) (the universe of discourse is the real
numbers). Its negation is: ∃x ∀y ¬(x < y), i.e., ∃x ∀y (x < y). (Note that among real
numbers x < y is equivalent to x ≥ y, but formally they are different predicates.)
Proofs
Answer : Assuming x > 2 and y > 3 and adding the inequalities term by term we
get: x + y > 2 + 3 = 5.
That is an example of direct proof. In a direct proof we assume the hypothesis together
with axioms and other theorems previously proved and we derive the conclusion from
them.
Answer : We must prove that x + y > 5 → (x > 2) ∨ (y > 3). An indirect proof
consists of proving ¬((x > 2) ∨ (y > 3)) → ¬(x + y > 5). In fact: ¬((x > 2) ∨ (y > 3))
is the same as (x ≤ 2) ∧(y ≤ 3), so adding both inequalities we get x + y ≤ 5, which is the
same as ¬(x + y > 5).
Answer : We assume the hypothesis x + y > 5. From here we must conclude that x >
2 or y > 3. Assume to the contrary that ―x > 2 or y > 3‖ is false, so x ≤ 2 and y ≤
3. Adding those inequalities we get
x ≤ 2 + 3 = 5, which contradicts the hypothesis x + y > 5. From here we conclude that
the assumption ―x ≤ 2 and y ≤ 3‖ cannot be right, so ―x > 2 or y > 3‖ must be true.
√
Exercise : Prove by contradiction that 2 is not a rational number, i.e., there are no
√
integers a, b such that 2 = a/b.
√ √
Answer : Assume that 2 is rational, i.e., 2 = a/b, where a and b are integers and the
fraction is written in least terms. Squaring both sides we have 2 = a2 /b2 , hence 2 b2 =
a2 . Since the left hand side is even, then a2 is even, but this implies that a itself is
even, so a = 2 a!. Hence: 2 b2 = 4 a!2 , and simplifying: b2 = 2 a! 2 . This implies that b2
is even, so b is even: b = 2b! . Consequently a/b = 2a! /2b! = a!/b!, contradicting the
hypothesis that a/b was in least terms.
p1
p2
.
pn
∴ q
or
p1 , p2 , . . . , pn / ∴ q
p→qp
∴ q
In order to check whether it is valid we must examine the following truth table:
p q p →p q
T T T q T T
T F F T F
F T T F T
F F T F F
If we look now at the rows in which both p → q and p are true (just the first row) we
see that also q is true, so the argument is valid.
Arguments are usually written using three columns. Each row con- tains a label, a
statement and the reason that justifies the introduction of that statement in the
argument. That justification can be one of the following:
Example : Consider the following statements: ―I take the bus or I walk. If I walk I
get tired. I do not get tired. Therefore I take the bus.‖ We can formalize this by
calling B = ―I take the bus‖, W = ―I walk‖ and T = ―I get tired‖. The premises are B
∨ W , W → T and
¬T , and the conclusion is B. The argument can be described in the
following steps:
step statement reason
1) W →T Premise
2) ¬T Premise
3) ¬W 1,2, Modus Tollens
4) B ∨W Premise
5) ∴B 4,3, Disjunctive Syllogism
2. Existential Instantiation. If ∃x p(x) is true, then p(a) is true for some specific
element a in the universe of discourse; i.e.:
∃x p(x)
∴ p(a)
The difference respect to the previous rule is the restriction in the meaning of a,
which now represents some (not any) element of the universe of discourse. So, for
instance, from ∃x (x2 = 2) (the universe of discourse is the real numbers) we derive
√ √
the existence of some element, which we may represent ± 2, such that (± 2)2 = 2.
4. Existential Generalization. If p(a) is true for some specific ele- ment a in the
universe of discourse, then ∃x p(x) is true; i.e.:
p(a)
∴ ∃x p(x)
For instance: from 7 + 1 = 8 we can derive ∃x (x + 1 = 8).
Mathematical Induction
The Well Ordering Principle
Mathematical Induction
Recursive Definitions
UNIT 4 6 Hours
INDUCTION PRINCIPLE:
The induction principle states as follows : let S(n) denote an open statement that involves a
positive integer n .suppose that the following conditions hold ;
1. S(1) is true
2. If whenever S(k) is true for some particular , but arbitrarily chosen k €Z+ , then
S(k+1) is true. Then S(n) is true for all n € Z+ . Z+ denotes the set of all positive
integers .
1) Basis step: verify that the statement S(1) is true ; i.e. verify that S(n) is true for n=1.
2) Induction step: assuming that S(k) is true , where k is an integer1, show that
S(k+1) is true.
The validity of the Principle of Mathematical Induction is obvious. The basis step
states that P (1) is true. Then the inductive step implies that P (2) is also true. By
the inductive step again we see that P (3) is true, and so on. Consequently the
property is true for all positive integers.
Remark : In the basis step we may replace 1 with some other integer m. Then the
conclusion is that the property is true for every integer n greater than or equal to m.
Example : Prove that the sum of the n first odd positive integers is
n2 , i.e., 1 + 3 + 5 + • • • + (2n — 1) ‗ n2 .
2. Inductive Step: Assume (Induction Hypothesis ) that the prop- erty is true for some
positive integer n, i.e.: S(n) ‗ n2 . We must prove that it is also true for n + 1, i.e., S(n
+ 1) ‗ (n + 1)2 . In fact:
S(n + 1) ‗ n2 + 2n + 1 ‗ (n + 1)2 .
This completes the induction, and shows that the property is true for all
positive integers.
Answer : This is an example in which the property is not true for all
positive integers but only for integers greater than or equal to 3.
This completes the induction, and shows that the property is true for all n ≥
3.
n (n + 1)
• 1 + 2 + 3 + ••• + n ‗
.
2
n (n + 1) (2n + 1)
• 1 + 2 + 3 + ••• + n ‗
2 2 2 2
.
6
• 13 + 23 + 33 + • • • + n3 ‗ (1 + 2 + 3 + • • • + n)2 .
n ‗ 2.
√
Hence starting with a fractional representation of 2 ‗ a/b we end up with another
√
fractional representation 2 ‗ b/a! with a smaller numerator b < a. Repeating the
same argument with the fraction b/a! we get another fraction with an even smaller
√
numerator, and so on. So the set of possible numerators of a fraction representing 2
cannot have a smallest element, contradicting the well-ordering principle.
√
Consequently, our assumption that 2 is rational has to be false.
Reccurence relations
Here we look at recursive definitions under a different point of view. Rather than
definitions they will be considered as equations that we must solve. The point is
that a recursive definition is actually a def- inition when there is one and only one
object satisfying it, i.e., when the equations involved in that definition have a
unique solution. Also, the solution to those equations may provide a closed-form
Dept. of CSE, SJBIT Page 43
DMS 10CS34
The basis of the recursive definition is also called initial conditions of the
recurrence. So, for instance, in the recursive definition of the Fibonacci sequence,
the recurrence is
Fm ‗ Fm−1 + Fm−2
or
Fm — Fm−1 — Fm−2 ‗ O ,
and the initial conditions are
F0 ‗ O, F1 ‗ 1 .
One way to solve some recurrence relations is by iteration, i.e., by using the
recurrence repeatedly until obtaining a explicit close-form formula. For instance
consider the following recurrence relation:
xm ‗ r xm−1 (n > O) ; x0 ‗ A .
xm ‗ r xm−1 ‗ r 2
xm−2 ‗
3 xm−3 ‗ • • • ‗ r m
x0 ‗ A rm ,
r
hence the solution is xm ‗ A rm .
xn ‗ r xn—1 (n > O) ; x0 ‗ A .
Its general solution is
xn ‗ A r n ,
which is a geometric sequence with ratio r.
xn ‗ r xn—1 + cn (n > O) ; x0 ‗ A .
Using the summation notation, its solution can be expressed like this:
n
X
xn ‗ A r n + ck rn—k .
k=1
xn ‗ A + c n if r ‗ 1 .
Example : Assume that a country with currently 1OO million people has a
population growth rate (birth rate minus death rate) of 1% per year, and it also
receives 1OO thousand immigrants per year (which are quickly assimilated and
reproduce at the same rate as the native population). Find its population in 1O
years from now. (Assume that all the immigrants arrive in a single batch at the
end of the year.)
OOO, So:
Dept. of CSE, SJBIT Page 45
DMS 10CS34
462, 317
.
xm ‗ xm—1 + c + d n (n > O) ; x0 ‗ A .
The solution is now
m
X d n (n + 1)
xm ‗ A + (c + d k) ‗ A + c n + .
k=1 2
C0 xm + C1 xm—1 + C2 xm—2 ‗ O .
C0 c rm + C1 c rm—1 + C2 c rm—2 ‗ O ,
C 0 r 2 + C1 r + C2 ‗ O .
Let r1 , r2 be the two (in general complex) roots of the above equation. They
are called characteristic roots. We distinguish three cases:
xm ‗ c1 rm + c2 rm ,
1 2
xm ‗ c1 rm + c2 n rm ,
RECURSIVE DEFINITIONS:
RECURRENCE RELATIONS:- The important methods to express the
recurrance formula in explict form are
1) BACKTRACKING METHOD
2) CHARACTERISTIC EQUATION METHOD
BACKTRACKING METHOD:
This is suitable method for linear non-homogenous recurrence relation of the type
xn= r x n-1 +s
The general method to find explicit formula
xn = r n-1 x1 + s(r n-1-1)/ (r-1) where r1 is the general explicit
CHARACTERISTIC EQUATION:
a n = r1 a n-1 + r 2 a n-2+…..+ r k a n-k..(1) is a LHRR of degree K .
x k = r1 x k-1 + r2 x k-2+….+r k is called characteristic equation.
• Let a n = r1 a n-1 + r2 a n-2 be LHRR of degree 2. its characteristic equation is x2 =
r1 x + r2 or x 2 –r1 x- r2=0. if the characteristic equation has 2 distinct roots e1 , e2
then the explicit formula of the recurrence relation in a n= u e n 1 + v e n 2 where u
and v depends on the initial values.
• Let an = r1 a n-1 + r2 an-2 be a LHRR of degree 2 . Its characteristic equation is x 2
–r1 x – r2 =0 if the characteristic equation has repeated roots e, then the explicit
formula is an =u e n + v n e n where u and v depends on the initial values.
PART – B
UNIT 5 7 Hours
Relations
SYLLABUS
Relations and Functions: Cartesian Products and Relations, Functions –
Plain and One-to-One, Onto Functions – Stirling Numbers of the Second
Kind, Special Functions, The Pigeon-hole Principle, Function
Composition and Inverse Functions
Introduction
Product set: If A and B are any 2 non-empty sets then the product set of A and B are the
Cartesian product or product of A and B.
AXB≠BXA
Then, A X B = {(1, a), (1, b), (2, a), (2, b), (3, a), (3, b)}
B X A = {(a, 1), (a, 2), (a, 3), (b, 1), (b, 2), (b, 3)}
AXB≠BxA
A X (B X C) = {(1, (a, x)), (1, (a, y)), (1, (b, x)), (1, (b, y)),
(2, (a, x)) (2, (a, y)), (2, (b, x)), (2, (b, y))}
(A X B) X C = {((1, a), x), ((1, a), y), ((1, b), x), ((1, b), y),
*Remarks:
a. A X (B X C) = (A X B) X C
b. A X A = A2
c. If R is the set of all real numbers then R x R = R2, set of all points in plane.
Partition set: Let 'A' be a non-empty set. A partition of 'A' or quotient set of 'A' is a
collection P of subsets of
Example:
Let,
A = {1, 2, 3, 4, 5} then,
Relations: Let A and B be any two non-empty sets. A relation R from a set A to the set B
is a subset of A x B.
Example:
Let,
A = {1, 2, 3, 4, 5}, Let R be a relation on A defined as a R b if a<b. R = {(1, 2), (1, 3), (1,
4), (1, 5) (2, 3), (2, 4), (2, 5), (3, 4), (3, 5), (4, 5)}
=> R A X A.
R(x) = {y ЄB/ x R y}
= U R(x) for x Є A1
Example:
Let,
A = {a, b, c, d}
R = {(a, a), (a, b), (b, c), (c, a) (c, b) (d, a)}
R (a) = {a, b}
R (b) = {c}
R (c) = {a, b}
R (d) = {a}
Let,
A1 = {a, c} be a subset of A,
= {a, b} U {a, b}
= {a, b}
Matrix of a relation / Relation Matrix: Let A = {a1, a2, a3………. am} and B = {b1, b2,
b3...bn} be any two finite sets.
Let R be relation from A to B then the matrix of the relation R is defined as the m x n
matrix,
MR= [Mij]
= 0, if (ai, bj) R
Example:
(a)Let,
10
Thus, Mr = 01
10
1001
(b) Given MR = 0110 . Find Relation R.
1010
Define set,
R = {(a1, b2) (a1, b4) (a2, b2) (a2, b3) (a3, b1) (a3, b3)}
(a)Draw a small circle for each element of A and label the circles with the corresponding
element of A. These circles are called "Vertices".
(c)The resulting picture representing the relation R is called the "directed graph of R" or
"digraph of R".
Example:
A = {1, 2, 3, 4}
R = {(1, 1), (1, 3), (2, 1), (2, 3), (3, 2), (4, 3)}
Diagram:
4
1
3
2
The "indegree" of a ЄA is the number of elements b Є A such that b R a.
1 2 2
2 1 2
3 3 1
4 0 1
A = {1, 2, 3, 4} B = {1, 4, 6, 8, 9}
10000
01000
Mr =
00001
00000
Properties of a relation:
Example: A = {I, 2, 3}
Therefore, R is reflexive.
A relation R on a set A is "non-reflexive" if ‗a‘ is not relation to ‗a‘ for some a ЄA or (a,
a) R for some a Є A
A = {1, 2, 3}
R = {(1, 1), (2, 1), (3, 2), (3, 3)} => (2, 2) R
Therefore, R is not-reflexive.
A relation R on a set A is ―not irreflexive‖ if ‗a‘ is not Relation to ‗a‘ for some a Є A.
Example: Let A = {1, 2, 3} and R = {(1, 1) (1, 2) (2, 1) (3, 2) (2, 3)}
Therefore, R is symmetric.
Therefore, R is asymmetric.
R is not symmetric.
Example: Let, A = {1, 2, 3} and R = {(1, 1), (1, 2), (3, 2)}
Example: Let, A = {1, 2, 3} and R = {(1, 1), (1, 3), (2, 3), (3, 1) (2, 1), (3, 3)} (all should
satisfy)
(a) Reflexive
(c) Transitive.
Symmetric: Let a R b
=> Є 1R2
Reflexive: a R a V a Є A
Symmetric: Let a R b
=> 1 R 3
=> 3 R 1
=> b R a
Hence, R is symmetric.
=> 1 R 2 and 2 R 3
=> 1 R 3
=> a R c
Hence, R is transitive
c. A = {1, 2, 3}
R = A x A = {(1, 1)(1, 2)(1, 3)(2, 1)(2, 2)(2, 3)(3, 1)(3, 2)(3, 3)}
OR
=> (a, a) Є R
=> a Є R (a)
=>a R x and b R x
This partition determines the relation R in the sense that a R b if a and b belong to the
same block of the partition.
Hence proved…..
Dept. of CSE, SJBIT Page 58
DMS 10CS34
Manipulation of relations:
a (R U S) b if either a R b or a S b
a (R n S) b if a R b and a S b
R2 =R o R = {(a, a), (a, c) (a, b) (b, a) (b, c) (b, b) (c, a) (c, b) (c, c)}
Reflexive closure: Let R be a relation on a set' A'. Suppose R lacks a particular property,
the smallest relation that contain R and which, processes the desired property is called the
closure of R with respective a property in question.
Given a relation R on a set' A' the relation R1 = (A U R) is the "reflexive closure of R‖.
Example:
A = {1, 2, 3}
Symmetric closure : If R is not symmetric then there exists (x, y) A such that (x, y) Є R,
but (y, x) Є R. To make R symmetric we need to add the ordered pairs of R-1.
A = {1, 2, 3}
R = {(1, 1) (1, 2) (2, 1) (1, 3) (3, 2)} find the symmetric closure of R.
Solution: We know that, R is not symmetric because (1, 3) Є R but (3, 1) Є R and (3, 2)
Є R but (2, 3) Є R.
Therefore, R1 = R U R-l = {(1, 1) (1, 2) (2, 1) (1, 3) (3, 1) (3, 2) (2, 3)}
Properties of Relations
Computer Recognition
Zero-One Matrices
Directed Graphs
Partial Orders
Hasse Diagrams
Equivalence Relations and
Partitions
UNIT 6 7 Hours
Functions
SYLLABUS
Relations contd.: Properties of Relations, Computer Recognition –
Zero-One Matrices and Directed Graphs, Partial Orders – Hasse
Diagrams, Equivalence Relations and Partitions
Introduction
A person counting students present in a class assigns a number to each student under
consideration. In this case a correspondence between two sets is established: between
students understand whole numbers. Such correspondence is called functions. Functions
are central to the study of physics and enumeration, but they occur in many other
situations as well. For instance, the correspondence between the data stored in computer
memory and the standard symbols a, b, c... z, 0, 1,...9,?,!, +... into strings of O's and I's
for digital processing and the subsequent decoding of the strings obtained: these are
functions. Thus, to understand the general use of functions, we must study their
properties in the general terms of set theory, which is will be we do in this chapter.
Definition: Let A and B be two sets. A function f from A to B is a rule that assigned to
each element x in A exactly one element y in B. It is denoted by
f: A B
Note:
Therefore, R is a function and S is not a function. Since the element 1has two images
band d, S is not a function.
Example: Let A = {1, 2, 3, 4} determine whether or not the following relations on A are
functions.
2. g={(3,1),(4,2),(1,1)}
g is a function
3. h={(2,1),(3,4),(1,4),(2,1),(4,4)}
h is a function
4. Let A= {0, ±1, ±2, 3}. Consider the function F: A R, where R is the set of all real
numbers, defined by f(x) =x3 -2x2 +3x+1 for xA. Find the range of f.
f (0) =1
f (1) =1-2+3+1=3
f (-1) =-1-2-3+1=-5
f (2) =8-8-6+1=7
f (3) =27-18+9+1=19
5. If A= {0, ±1, ±2} and f: A R is defined by f(x) =x2-x+1, xA find the range.
f (0) =1
f (1) =1-1+1=1
f (-1) =1+1+1=3
f (2) =4-2+1=3
f (-2) =4+2+1=7
Range = {1, 3, 7}
Types of functions:
1. Everywhere defined -2
A function f: A ~ B is everywhere defined if domain of f equal to A (dom f
= A)
Example:
5. Bijection function
A function f: A B is Bijection if it is both onto and one-to-one.
6. Invertible function
Example: f f-1
A f(a)
B B A
a
f(b)
b f--1(f(a)) a
f(c)
c f--1(f(b)) b
f(d)
d f--1(f(c)) c
f--1(f(d)) d
Here f: A B f -1 B A
A = {a1, a2, a3} B = {b1, b2, b3} C = { c1, c2} D = {d1, d2, d3, d4}
A function f: A~ A such that f (a) = a, 'if a Є A is called the identity function or identity
mapping on A. Dom (t) = Ran (t) = A
Constant function
A function f: A B such that f (a) =c, a dom (f) where c is a fixed element of B, is
called a constant function.
Into function
One-to-one correspondence
Composition of function
Inverse function
Consider a function f: A (B. Then f is a relation from A to B with Dom (f) (A and Ran (f)
(B. Its inverse, f -1, is a relation from B to A which is such that if whenever (a, b) (f then
(b, a) (f -1)
(f -1) -1 = f
Definition
Also, (f -1) -1 = f
Characteristic function
Introduction
Definition
fA (x) = 1 if x A
0 if x A
Proof:
i. Let x AUB then f AUB (x) = 1. But if x AUB then there are three cases
case1: let x A but not in B then fA (x) = 1 and fB (x)=0 f AUB (x) = 1 = f A (x)
+ f A (x) - f A (x) . f B (x)
[Because 1+0+0]
f (A B) C (x) = f A (B C) (x) x
LHS = f (A B) C
= f (D C) where D = A B
= f D+ f c–2f Df c
= f c + f D (1-2 f c )
= f D + f A B (1-2 f c)
= f c + (f A + f B - 2 f A f B) (1 – 2 f c)
= f c+ f A+ f B- 2 f A f B- 2 f A f C- 2 f B f C+ 4 f A f B f C
= f A + (f B + f c – 2 f B f C) – 2 f A (f B + f C – 2 f B f C)
= f A + f B + f C - 2 f B f C (1- 2 f A)
= f A + f BC (1 - 2 f A)
= f A + f BC - 2 f A f BC
= f A (BC)
= RHS
(A B) C = A (B C)
Permutation function
A permutation on ‗A‘ is a bijection of ‗A‘ onto itself. Let ‗A‘ = {a1, a2, a3, -----------
an}.Where A is a finite set, if P is a permutation on A then P can be represented as ,
P= a1 a2 a3-----------an
In this permutation suppose r elements of A say {b1, b2, b3, -------- br} are such that P
(b1) =b2, P (b2) =b3, .....P(br-1 ) =br , P(br ) =b1 , and the remaining elements of A are
images of themselves, Then P is called a cycle of length ‗r‘ , and is denoted by (b1, b2,
b3 ……. br).
Example 1:
3 2 4 1
P is a cycle of length 3.
Example 2:
A = {1, 2, 3, 4, 5, 6} and P = 1 2 3 4 5 6
3 2 4 6 5 1
P is a cycle of length 4.
Transposition
A cycle of length 2 is called a ―transposition‖ if A = {a1, a2, a3, ---- an} then P = (ai, aj),
i j is a transposition of A.
Example:
A = {1, 2, 3, 4, 5, 6} compute
2. (5, 6, 3) o (4, 1, 3, 5)
P1 = 1 2 3 4 5 6
3 2 5 1 4 6
P2 = (5, 6, 3) = 1 2 3 4 5 6
1 2 5 4 6 3
P10P2 =
1 2 3 4 5 6 1 2 3 4 5 6
3 2 5 1 4 6 1 2 5 4 6 3
2. P2 0 P1 = (5, 6, 3) o (4, 1, 3, 5) = 1 1 2 3 4 5 6
5 2 6 1 4 3
Example:
1 1 2 3 4 5 6 7 8
1. P =
4 2 1 6 5 8 7 3
P = (1, 3) o (1, 8) o (1, 6) o (1, 4)
P is an even permutation.
Hashing function
Introduction
Suppose we want to store the mail address of all voters of a large city in n number of
files, numbered from 0 to n-1 , in such a way that the file containing the address any
chosen voter can be located almost instantly. The following is one way of doing this task
First, to each voter let us assign a unique positive integer as an identification number.
Next, to each identification number, let us assign a unique positive integer called a key.
The keys can be such that two identification numbers can have the same key but two
different keys are not assigned to the same identification number.
Therefore the number of identification number will be equal to the number of voters , but
the number, of keys can be less than the no. of identification number.
Definition
Let A denote the set of all keys and B = {0, 1, 2, ------- (n-1)} , the set of all files.
The function hn is called hashing function. For this function a set of all keys is domain.
NOTE: The key need not be different from the identification number. If the keys are
identical with the identification number, then the domain of the hashing function is the
set of all identification number.
Definitions,
Examples,
Elementary Properties,
Homomorphisms,
Isomorphisms, and
Cyclic Groups,
Cosets, and
Lagrange‘s Theorem.
UNIT 7 6 hrs
GROUPS
SYLLABUS
Groups: Definitions, Examples, and Elementary Properties,
Homomorphisms, Isomorphisms, and Cyclic Groups, Cosets, and
Lagrange’s Theorem
In mathematics, a discrete group is a group G equipped with the discrete topology. With
this topology G becomes a topological group. A discrete subgroup of a topological
group G is a subgroup H whose relative topology is the discrete one. For example, the
integers, Z, form a discrete subgroup of the reals, R, but the rational numbers, Q, do not.
Any group can be given the discrete topology. Since every map from a discrete space is
continuous, the topological homomorphisms between discrete groups are exactly the
group homomorphisms between the underlying groups. Hence, there is an isomorphism
between the category of groups and the category of discrete groups. Discrete groups can
therefore be identified with their underlying (non-topological) groups. With this in mind,
the term discrete group theory is used to refer to the study of groups without topological
structure, in contradistinction to topological or Lie group theory. It is divided, logically
but also technically, into finite group theory, and infinite group theory.
There are some occasions when a topological group or Lie group is usefully endowed
with the discrete topology, 'against nature'. This happens for example in the theory of the
Bohr compactification, and in group cohomology theory of Lie groups.
Properties:
Since topological groups are homogeneous, one need only look at a single point to
determine if the group is discrete. In particular, a topological group is discrete if and only
if the singleton containing the identity is an open set.
A discrete group is the same thing as a zero-dimensional Lie group (uncountable discrete
groups are not second-countable so authors who require Lie groups to satisfy this axiom
do not regard these groups as Lie groups). The identity component of a discrete group is
just the trivial subgroup while the group of components is isomorphic to the group itself.
Since the only Hausdorff topology on a finite set is the discrete one, a finite Hausdorff
topological group must necessarily be discrete. It follows that every finite subgroup of a
Hausdorff group is discrete.
Discrete normal subgroups play an important role in the theory of covering groups and
locally isomorphic groups. A discrete normal subgroup of a connected group G
necessarily lies in the center of G and is therefore abelian.
Other properties:
Examples:
Frieze groups and wallpaper groups are discrete subgroups of the isometry group
of the Euclidean plane. Wallpaper groups are cocompact, but Frieze groups are
not.
A space group is a discrete subgroup of the isometry group of Euclidean space of
some dimension.
A crystallographic group usually means a cocompact, discrete subgroup of the
isometries of some Euclidean space. Sometimes, however, a crystallographic
group can be a cocompact discrete subgroup of a nilpotent or solvable Lie group.
Every triangle group T is a discrete subgroup of the isometry group of the sphere
(when T is finite), the Euclidean plane (when T has a Z + Z subgroup of finite
index), or the hyperbolic plane.
Group homomorphism:
Image of a Group homomorphism(h) from G(left) to H(right). The smaller oval inside H
is the image of h. N is the kernel of h and aN is a coset of h.
In mathematics, given two groups (G, *) and (H, ·), a group homomorphism from (G, *)
to (H, ·) is a function h : G → H such that for all u and v in G it holds that
where the group operation on the left hand side of the equation is that of G and on the
right hand side that of H.
From this property, one can deduce that h maps the identity element eG of G to the
identity element eH of H, and it also maps inverses to inverses in the sense that
h(u − 1) = h(u) − 1.
Hence one can say that h "is compatible with the group structure".
Older notations for the homomorphism h(x) may be xh, though this may be confused as an
index or a general subscript. A more recent trend is to write group homomorphisms on
the right of their arguments, omitting brackets, so that h(x) becomes simply x h. This
approach is especially prevalent in areas of group theory where automata play a role,
since it accords better with the convention that automata read words from left to right.
In areas of mathematics where one considers groups endowed with additional structure, a
homomorphism sometimes means a map which respects not only the group structure (as
above) but also the extra structure. For example, a homomorphism of topological groups
is often required to be continuous.
If the homomorphism h is a bijection, then one can show that its inverse is also a group
homomorphism, and h is called a group isomorphism; in this case, the groups G and H
are called isomorphic: they differ only in the notation of their elements and are identical
for all practical purposes.
automorphism group of (Z, +) contains only two elements, the identity transformation
and multiplication with -1; it is isomorphic to Z/2Z.
If G and H are abelian (i.e. commutative) groups, then the set Hom(G, H) of all group
homomorphisms from G to H is itself an abelian group: the sum h + k of two
homomorphisms is defined by
(h + k) o f = (h o f) + (k o f) and g o (h + k) = (g o h) + (g o k).
This shows that the set End(G) of all endomorphisms of an abelian group forms a ring,
the endomorphism ring of G. For example, the endomorphism ring of the abelian group
consisting of the direct sum of m copies of Z/nZ is isomorphic to the ring of m-by-m
matrices with entries in Z/nZ. The above compatibility also shows that the category of all
abelian groups with group homomorphisms forms a preadditive category; the existence of
direct sums and well-behaved kernels makes this category the prototypical example of an
abelian category.
Cyclic group
In group theory, a cyclic group is a group that can be generated by a single element, in
the sense that the group has an element g (called a "generator" of the group) such that,
when written multiplicatively, every element of the group is a power of g (a multiple of g
when the notation is additive).
Definition
The six 6th complex roots of unity form a cyclic group under multiplication. z is a primitive
element, but z2 is not, because the odd powers of z are not a power of z2.
For example, if G = { g0, g1, g2, g3, g4, g5 } is a group, then g6 = g0, and G is cyclic. In
fact, G is essentially the same as (that is, isomorphic to) the set { 0, 1, 2, 3, 4, 5 } with
addition modulo 6. For example, 1 + 2 = 3 (mod 6) corresponds to g1·g2 = g3, and 2 + 5 =
1 (mod 6) corresponds to g2·g5 = g7 = g1, and so on. One can use the isomorphism φ
defined by φ(gi) = i.
For every positive integer n there is exactly one cyclic group (up to isomorphism) whose
order is n, and there is exactly one infinite cyclic group (the integers under addition).
Hence, the cyclic groups are the simplest groups and they are completely classified.
The name "cyclic" may be misleading: it is possible to generate infinitely many elements
and not form any literal cycles; that is, every gn is distinct. (It can be said that it has one
infinitely long cycle.) A group generated in this way is called an infinite cyclic group,
and is isomorphic to the additive group of integers Z.
Furthermore, the circle group (whose elements are uncountable) is not a cyclic group—a
cyclic group always has countable elements.
Since the cyclic groups are abelian, they are often written additively and denoted Zn.
However, this notation can be problematic for number theorists because it conflicts with
the usual notation for p-adic number rings or localization at a prime ideal. The quotient
notations Z/nZ, Z/n, and Z/(n) are standard alternatives. We adopt the first of these here
to avoid the collision of notation. See also the section Subgroups and notation below.
One may write the group multiplicatively, and denote it by Cn, where n is the order
(which can be ∞). For example, g3g4 = g2 in C5, whereas 3 + 4 = 2 in Z/5Z.
Properties
The fundamental theorem of cyclic groups states that if G is a cyclic group of order n
then every subgroup of G is cyclic. Moreover, the order of any subgroup of G is a divisor
of n and for each positive divisor k of n the group G has exactly one subgroup of order k.
This property characterizes finite cyclic groups: a group of order n is cyclic if and only if
for every divisor d of n the group has at most one subgroup of order d. Sometimes the
equivalent statement is used: a group of order n is cyclic if and only if for every divisor d
of n the group has exactly one subgroup of order d.
Every finite cyclic group is isomorphic to the group { [0], [1], [2], ..., [n − 1] } of integers
modulo n under addition, and any infinite cyclic group is isomorphic to Z (the set of all
integers) under addition. Thus, one only needs to look at such groups to understand the
properties of cyclic groups in general. Hence, cyclic groups are one of the simplest
groups to study and a number of nice properties are known.
G is abelian; that is, their group operation is commutative: gh = hg (for all h in G). This is
so since g + h mod n = h + g mod n.
If n is finite, then gn = g0 is the identity element of the group, since kn mod n = 0 for any
integer k.
If n = ∞, then there are exactly two elements that generate the group on their own:
namely 1 and −1 for Z
If n is finite, then there are exactly φ(n) elements that generate the group on their own,
where φ is the Euler totient function
Every subgroup of G is cyclic. Indeed, each finite subgroup of G is a group of { 0, 1, 2, 3,
... m − 1} with addition modulo m. And each infinite subgroup of G is mZ for some m,
which is bijective to (so isomorphic to) Z.
Gn is isomorphic to Z/nZ (factor group of Z over nZ) since Z/nZ = {0 + nZ, 1 + nZ, 2 +
nZ, 3 + nZ, 4 + nZ, ..., n − 1 + nZ} { 0, 1, 2, 3, 4, ..., n − 1} under addition modulo n.
More generally, if d is a divisor of n, then the number of elements in Z/n which have
order d is φ(d). The order of the residue class of m is n / gcd(n,m).
If p is a prime number, then the only group (up to isomorphism) with p elements is the
cyclic group Cp or Z/pZ.
The direct product of two cyclic groups Z/nZ and Z/mZ is cyclic if and only if n and m
are coprime. Thus e.g. Z/12Z is the direct product of Z/3Z and Z/4Z, but not the direct
product of Z/6Z and Z/2Z.
The definition immediately implies that cyclic groups have very simple group
presentation C∞ = < x | > and Cn = < x | xn > for finite n.
A primary cyclic group is a group of the form Z/pk where p is a prime number. The
fundamental theorem of abelian groups states that every finitely generated abelian group
is the direct product of finitely many finite primary cyclic and infinite cyclic groups.
Z/nZ and Z are also commutative rings. If p is a prime, then Z/pZ is a finite field, also
denoted by Fp or GF(p). Every field with p elements is isomorphic to this one.
The units of the ring Z/nZ are the numbers coprime to n. They form a group under
multiplication modulo n with φ(n) elements (see above). It is written as (Z/nZ)×. For
example, when n = 6, we get (Z/nZ)× = {1,5}. When n = 8, we get (Z/nZ)× = {1,3,5,7}.
The group (Z/pZ)× is cyclic with p − 1 elements for every prime p, and is also written
(Z/pZ)* because it consists of the non-zero elements. More generally, every finite
subgroup of the multiplicative group of any field is cyclic.
Examples
In 2D and 3D the symmetry group for n-fold rotational symmetry is Cn, of abstract group
type Zn. In 3D there are also other symmetry groups which are algebraically the same, see
Symmetry groups in 3D that are cyclic as abstract group.
Note that the group S1 of all rotations of a circle (the circle group) is not cyclic, since it is
not even countable.
The nth roots of unity form a cyclic group of order n under multiplication. e.g., 0 = z3 − 1
= (z − s0)(z − s1)(z − s2) where si = e2πi / 3 and a group of {s0,s1,s2} under multiplication is
cyclic.
The Galois group of every finite field extension of a finite field is finite and cyclic;
conversely, given a finite field F and a finite cyclic group G, there is a finite field
extension of F whose Galois group is G.
Representation
The cycle graphs of finite cyclic groups are all n-sided polygons with the elements at the
vertices. The dark vertex in the cycle graphs below stand for the identity element, and the
other vertices are the other elements of the group. A cycle consists of successive powers
of either of the elements connected to the identity element.
C1 C2 C3 C4 C5 C6 C7 C8
The representation theory of the cyclic group is a critical base case for the representation
theory of more general finite groups. In the complex case, a representation of a cyclic
group decomposes into a direct sum of linear characters, making the connection between
character theory and representation theory transparent. In the positive characteristic case,
the indecomposable representations of the cyclic group form a model and inductive basis
for the representation theory of groups with cyclic Sylow subgroups and more generally
the representation theory of blocks of cyclic defect.
All subgroups and quotient groups of cyclic groups are cyclic. Specifically, all subgroups
of Z are of the form mZ, with m an integer ≥0. All of these subgroups are different, and
apart from the trivial group (for m=0) all are isomorphic to Z. The lattice of subgroups of
Z is isomorphic to the dual of the lattice of natural numbers ordered by divisibility. All
factor groups of Z are finite, except for the trivial exception Z/{0} = Z/0Z. For every
positive divisor d of n, the quotient group Z/nZ has precisely one subgroup of order d, the
one generated by the residue class of n/d. There are no other subgroups. The lattice of
subgroups is thus isomorphic to the set of divisors of n, ordered by divisibility. In
particular, a cyclic group is simple if and only if its order (the number of its elements) is
prime.
Using the quotient group formalism, Z/nZ is a standard notation for the additive cyclic
group with n elements. In ring terminology, the subgroup nZ is also the ideal (n), so the
quotient can also be written Z/(n) or Z/n without abuse of notation. These alternatives do
not conflict with the notation for the p-adic integers. The last form is very common in
informal calculations; it has the additional advantage that it reads the same way that the
group or ring is often described verbally, "Zee mod en".
Endomorphisms
The endomorphism ring of the abelian group Z/nZ is isomorphic to Z/nZ itself as a ring.
Under this isomorphism, the number r corresponds to the endomorphism of Z/nZ that
maps each element to the sum of r copies of it. This is a bijection if and only if r is
coprime with n, so the automorphism group of Z/nZ is isomorphic to the unit group
(Z/nZ)× (see above).
Similarly, the endomorphism ring of the additive group Z is isomorphic to the ring Z. Its
automorphism group is isomorphic to the group of units of the ring Z, i.e. to {−1, +1}
C2 .
A group is called virtually cyclic if it contains a cyclic subgroup of finite index (the
number of cosets that the subgroup has). In other words, any element in a virtually cyclic
group can be arrived at by applying a member of the cyclic subgroup to a member in a
certain finite set. Every cyclic group is virtually cyclic, as is every finite group. It is
known that a finitely generated discrete group with exactly two ends is virtually cyclic
(for instance the product of Z/n and Z). Every abelian subgroup of a Gromov hyperbolic
group is virtually cyclic.
Group isomorphism
In abstract algebra, a group isomorphism is a function between two groups that sets up a
one-to-one correspondence between the elements of the groups in a way that respects the
given group operations. If there exists an isomorphism between two groups, then the
groups are called isomorphic. From the standpoint of group theory, isomorphic groups
have the same properties and need not be distinguished.
Given two groups (G, *) and (H, ), a group isomorphism from (G, *) to (H, ) is a
bijective group homomorphism from G to H. Spelled out, this means that a group
isomorphism is a bijective function such that for all u and v in G it holds
that
The two groups (G, *) and (H, ) are isomorphic if an isomorphism exists. This is
written:
Often shorter and more simple notations can be used. Often there is no ambiguity about
the group operation, and it can be omitted:
Sometimes one can even simply write G = H. Whether such a notation is possible without
confusion or ambiguity depends on context. For example, the equals sign is not very
suitable when the groups are both subgroups of the same group. See also the examples.
Conversely, given a group (G, *), a set H, and a bijection , we can make H
a group (H, ) by defining
Intuitively, group theorists view two isomorphic groups as follows: For every element g
of a group G, there exists an element h of H such that h 'behaves in the same way' as g
(operates with other elements of the group in the same way as g). For instance, if g
generates G, then so does h. This implies in particular that G and H are in bijective
correspondence. So the definition of an isomorphism is quite natural.
Examples
The group of all real numbers with addition, ( ,+), is isomorphic to the group of all
positive real numbers with multiplication ( +,×):
f(x) = ex
The group of integers (with addition) is a subgroup of , and the factor group /
is isomorphic to the group S1 of complex numbers of absolute value 1 (with
multiplication):
An isomorphism is given by
for every x in .
Generalizing this, for all odd n, Dih2n is isomorphic with the direct product of Dihn and
Z2.
If (G, *) is an infinite cyclic group, then (G, *) is isomorphic to the integers (with the
addition operation). From an algebraic point of view, this means that the set of all
integers (with the addition operation) is the 'only' infinite cyclic group.
Some groups can be proven to be isomorphic, relying on the axiom of choice, while it is
even theoretically impossible to construct concrete isomorphisms. Examples:
Properties
The kernel of an isomorphism from (G, *) to (H, ), is always {eG} where eG is the
identity of the group (G, *)
If (G, *) is a locally finite group that is isomorphic to (H, ), then (H, ) is also locally
finite.
The previous examples illustrate that 'group properties' are always preserved by
isomorphisms.
Cyclic groups
Let G be a cyclic group and n be the order of G. G is then the group generated by < x > =
{e,x,...,xn − 1}. We will show that
Define
, so that . Clearly, is
bijective.
Then
Consequences
From the definition, it follows that any isomorphism will map the identity
element of G to the identity element of H,
f(eG) = eH
for all u in G, and that the inverse map is also a group isomorphism.
The relation "being isomorphic" satisfies all the axioms of an equivalence relation. If f is
an isomorphism between two groups G and H, then everything that is true about G that is
only related to the group structure can be translated via f into a true ditto statement about
H, and vice versa.
Automorphisms
An automorphism always maps the identity to itself. The image under an automorphism
of a conjugacy class is always a conjugacy class (the same or another). The image of an
element has the same order as that element.
For all Abelian groups there is at least the automorphism that replaces the group elements
by their inverses. However, in groups where all elements are equal to their inverse this is
Dept. of CSE, SJBIT Page 89
DMS 10CS34
the trivial automorphism, e.g. in the Klein four-group. For that group all permutations of
the three non-identity elements are automorphisms, so the automorphism group is
isomorphic to S3 and Dih3.
In Zp for a prime number p, one non-identity element can be replaced by any other, with
corresponding changes in the other elements. The automorphism group is isomorphic to
Zp − 1. For example, for n = 7, multiplying all elements of Z7 by 3, modulo 7, is an
automorphism of order 6 in the automorphism group, because 36 = 1 ( modulo 7 ), while
lower powers do not give 1. Thus this automorphism generates Z6. There is one more
automorphism with this property: multiplying all elements of Z7 by 5, modulo 7.
Therefore, these two correspond to the elements 1 and 5 of Z6, in that order or
conversely.
The automorphism group of Z6 is isomorphic to Z2, because only each of the two
elements 1 and 5 generate Z6, so apart from the identity we can only interchange these.
For Abelian groups all automorphisms except the trivial one are called outer
automorphisms.
Non-Abelian groups have a non-trivial inner automorphism group, and possibly also
outer automorphisms.
The first, source encoding, attempts to compress the data from a source in order to
transmit it more efficiently. This practice is found every day on the Internet where the
common "Zip" data compression is used to reduce the network load and make files
smaller. The second, channel encoding, adds extra data bits to make the transmission of
data more robust to disturbances present on the transmission channel. The ordinary user
may not be aware of many applications using channel coding. A typical music CD uses
the Reed-Solomon code to correct for scratches and dust. In this application the
transmission channel is the CD itself. Cell phones also use coding techniques to correct
for the fading and noise of high frequency radio transmission. Data modems, telephone
transmissions, and NASA all employ channel coding techniques to get the bits through,
for example the turbo code and LDPC codes.
3-bit binary cube for finding Two example distances: 100->011 has distance 3 (red path);
Hamming distance 010->111 has distance 2 (blue path)
Two example distances: 0100->1001 has distance 3 (red path); 0110->1110 has distance 1 (blue
path)
In information theory, the Hamming distance between two strings of equal length is the
number of positions at which the corresponding symbols are different. Put another way, it
measures the minimum number of substitutions required to change one string into the
other, or the number of errors that transformed one string into the other.
Examples
Special properties
For a fixed length n, the Hamming distance is a metric on the vector space of the words
of that length, as it obviously fulfills the conditions of non-negativity, identity of
indiscernibles and symmetry, and it can be shown easily by complete induction that it
satisfies the triangle inequality as well. The Hamming distance between two words a and
b can also be seen as the Hamming weight of a−b for an appropriate choice of the −
operator.
For binary strings a and b the Hamming distance is equal to the number of ones in a
XOR b. The metric space of length-n binary strings, with the Hamming distance, is
known as the Hamming cube; it is equivalent as a metric space to the set of distances
between vertices in a hypercube graph. One can also view a binary string of length n as a
vector in Rn by treating each symbol in the string as a real coordinate; with this
embedding, the strings form the vertices of an n-dimensional hypercube, and the
Hamming distance of the strings is equivalent to the Manhattan distance between the
vertices.
The Hamming distance is named after Richard Hamming, who introduced it in his
fundamental paper on Hamming codes Error detecting and error correcting codes in
1950.[1] It is used in telecommunication to count the number of flipped bits in a fixed-
length binary word as an estimate of error, and therefore is sometimes called the signal
distance. Hamming weight analysis of bits is used in several disciplines including
information theory, coding theory, and cryptography. However, for comparing strings of
different lengths, or strings where not just substitutions but also insertions or deletions
have to be expected, a more sophisticated metric like the Levenshtein distance is more
appropriate. For q-ary strings over an alphabet of size q ≥ 2 the Hamming distance is
applied in case of orthogonal modulation, while the Lee distance is used for phase
modulation. If q = 2 or q = 3 both distances coincide.
On a grid (such as a chessboard), the points at a Lee distance of 1 constitute the von
Neumann neighborhood of that point.
Algorithm example
The Python function hamming_distance() computes the Hamming distance between two
strings (or other iterable objects) of equal length, by creating a sequence of zero and one
values indicating mismatches and matches between corresponding positions in the two
inputs, and then summing the sequence.
The following C function will compute the Hamming distance of two integers
(considered as binary values, that is, as sequences of bits). The running time of this
procedure is proportional to the Hamming distance rather than to the number of bits in
the inputs. It computes the bitwise exclusive or of the two inputs, and then finds the
Hamming weight of the result (the number of nonzero bits) using an algorithm of Wegner
(1960) that repeatedly finds and clears the lowest-order nonzero bit.
the Hamming metric. The weight of a vector (in the Hamming metric) is .
Note that
(3.1)
for any vectors (or, more generally, any vectors in a linear code). Using this, it is
easy to show that satisfies the properties of a metric:
, for all .
, for all .
This is called the ball of radius about . Since is finite, this ball has only a finite
number of elements. It is not hard to count them using a little bit of basic combinitorics. Since
this count shall be needed later, we record it in the following result.
proof: Let
This is called the shell of radius about . It is consists of all vectors with exactly
coordinates different from . There are ways to choose out of coordinates. There
is called the ISBN code. This is an -ary linear code of length . This is the same code used
in book numbering except that the number is denoted by on the inside cover of a book.
Generator matrix
In coding theory, a generator matrix is a basis for a linear code, generating all its
possible codewords. If the matrix is G and the linear code is C,
w=cG
where w is a unique codeword of the linear code C, c is a unique row vector, and a
bijection exists between w and c. A generator matrix for a (n, M = qk, d)q-code is of
dimension k×n. Here n is the length of a codeword, k is the number of information bits, d
is the minimum distance of the code, and q is the number of symbols in the alphabet
(thus, q = 2 indicates a binary code, etc.). Note that the number of redundant bits is
denoted r = n - k.
A generator matrix can be used to construct the parity check matrix for a code (and vice-
versa).
Equivalent Codes
Codes C1 and C2 are equivalent (denoted C1 ~ C2) if one code can be created from the
other via the following two transformations:
The generator matrices of equivalent codes can be obtained from one another via the
following transformations:
1. permute rows
2. scale rows
3. add rows
4. permute columns, and
5. scale columns.
Parity-check matrix
The rows of a parity check matrix are parity checks on the codewords of a code. That is,
they show how linear combinations of certain digits of each codeword equal zero. For
example, the parity check matrix
specifies that for each codeword, digits 1 and 2 should sum to zero and digits 3 and 4
should sum to zero.
The parity check matrix for a given code can be derived from its generator matrix (and
vice-versa). If the generator matrix for an [n,k]-code is in standard form
because
GHT = P − P = 0.
Negation is performed in the finite field mod q. Note that if the characteristic of the
underlying field is 2 (i.e., 1 + 1 = 0 in that field), as in binary codes, then − P = P, so the
negation is unnecessary.
For any valid codeword x, Hx = 0. For any invalid codeword , the syndrome S satisfies
.
Parity check
If no error occurs during transmission, then the received codeword r is identical to the
transmitted codeword x:
The receiver multiplies H and r to obtain the syndrome vector , which indicates
whether an error has occurred, and if so, for which codeword bit. Performing this
multiplication (again, entries modulo 2):
Since the syndrome z is the null vector, the receiver can conclude that no error has
occurred. This conclusion is based on the observation that when the data vector is
multiplied by , a change of basis occurs into a vector subspace that is the kernel of .
As long as nothing happens during transmission, will remain in the kernel of and the
multiplication will yield the null vector.
Coset
Only when H is normal will the right and left cosets of H coincide, which is one
definition of normality of a subgroup.
A coset is a left or right coset of some subgroup in G. Since Hg = g ( g−1Hg ), the right
cosets Hg (of H ) and the left cosets g ( g−1Hg ) (of the conjugate subgroup g−1Hg ) are the
same. Hence it is not meaningful to speak of a coset as being left or right unless one first
specifies the underlying subgroup.
For abelian groups or groups written additively, the notation used changes to g+H and
H+g respectively.
Examples
0 + H = {0, 2} = H
1 + H = {1, 3}
2 + H = {2, 0} = H
3 + H = {3, 1}.
There are therefore two distinct cosets, H itself, and 1 + H = 3 + H. Note that every
element of G is either in H or in 1 + H, that is, H ∪ (1 + H ) = G, so the distinct cosets of
H in G partition G. Since Z4 is an abelian group, the right cosets will be the same as the
left.
Another example of a coset comes from the theory of vector spaces. The elements
(vectors) of a vector space form an Abelian group under vector addition. It is not hard to
show that subspaces of a vector space are subgroups of this group. For a vector space V, a
subspace W, and a fixed vector a in V, the sets
are called affine subspaces, and are cosets (both left and right, since the group is
Abelian). In terms of geometric vectors, these affine subspaces are all the "lines" or
"planes" parallel to the subspace, which is a line or plane going through the origin.
General properties
Any two left cosets of H in G are either identical or disjoint — i.e., the left cosets form a
partition of G such that every element of G belongs to one and only one left coset.[1] In
particular the identity is in precisely one coset, and that coset is H itself; this is also the
only coset that is a subgroup. We can see this clearly in the above examples.
The left cosets of H in G are the equivalence classes under the equivalence relation on G
given by x ~ y if and only if x -1y ∈ H. Similar statements are also true for right cosets.
Index of a subgroup
All left cosets and all right cosets have the same order (number of elements, or cardinality
in the case of an infinite H), equal to the order of H (because H is itself a coset).
Furthermore, the number of left cosets is equal to the number of right cosets and is
known as the index of H in G, written as [G : H ]. Lagrange's theorem allows us to
compute the index in the case where G and H are finite, as per the formula:
This equation also holds in the case where the groups are infinite, although the meaning
may be less clear.
If H is not normal in G, then its left cosets are different from its right cosets. That is, there
is an a in G such that no element b satisfies aH = Hb. This means that the partition of G
into the left cosets of H is a different partition than the partition of G into right cosets of
H. (It is important to note that some cosets may coincide. For example, if a is in the
center of G, then aH = Ha.)
On the other hand, the subgroup N is normal if and only if gN = Ng for all g in G. In this
case, the set of all cosets form a group called the quotient group G /N with the operation ∗
defined by (aN )∗(bN ) = abN. Since every right coset is a left coset, there is no need to
differentiate "left cosets" from "right cosets".
Lagrange's theorem, in the mathematics of group theory, states that for any finite group
G, the order (number of elements) of every subgroup H of G divides the order of G. The
theorem is named after Joseph Lagrange.
This can be shown using the concept of left cosets of H in G. The left cosets are the
equivalence classes of a certain equivalence relation on G and therefore form a partition
of G. Specifically, x and y in G are related if and only if there exists h in H such that x =
yh. If we can show that all cosets of H have the same number of elements, then each coset
Dept. of CSE, SJBIT Page 101
DMS 10CS34
of H has precisely |H| elements. We are then done since the order of H times the number
of cosets is equal to the number of elements in G, thereby proving that the order H
divides the order of G. Now, if aH and bH are two left cosets of H, we can define a map
f : aH → bH by setting f(x) = ba-1x. This map is bijective because its inverse is given by
f −1(y) = ab−1y.
This proof also shows that the quotient of the orders |G| / |H| is equal to the index [G : H]
(the number of left cosets of H in G). If we write this statement as
|G| = [G : H] · |H|,
then, seen as a statement about cardinal numbers, it is equivalent to the Axiom of choice.
A consequence of the theorem is that the order of any element a of a finite group (i.e. the
smallest positive integer number k with ak = e, where e is the identity element of the
group) divides the order of that group, since the order of a is equal to the order of the
cyclic subgroup generated by a. If the group has n elements, it follows
an = e.
This can be used to prove Fermat's little theorem and its generalization, Euler's theorem.
These special cases were known long before the general theorem was proved.
The theorem also shows that any group of prime order is cyclic and simple.
Lagrange's theorem raises the converse question as to whether every divisor of the order
of a group is the order of some subgroup. This does not hold in general: given a finite
group G and a divisor d of |G|, there does not necessarily exist a subgroup of G with order
d. The smallest example is the alternating group G = A4 which has 12 elements but no
subgroup of order 6. A CLT group is a finite group with the property that for every
divisor of the order of the group, there is a subgroup of that order. It is known that a CLT
group must be solvable and that every supersolvable group is a CLT group: however
there exists solvable groups which are not CLT and CLT groups which are not
supersolvable.
There are partial converses to Lagrange's theorem. For general groups, Cauchy's theorem
guarantees the existence of an element, and hence of a cyclic subgroup, of order any
prime dividing the group order; Sylow's theorem extends this to the existence of a
subgroup of order equal to the maximal power of any prime dividing the group order. For
solvable groups, Hall's theorems assert the existence of a subgroup of order equal to any
unitary divisor of the group order (that is, a divisor coprime to its cofactor).
History
Lagrange did not prove Lagrange's theorem in its general form. He stated, in his article
Réflexions sur la résolution algébrique des équations,[1] that if a polynomial in n
variables has its variables permuted in all n ! ways, the number of different polynomials
that are obtained is always a factor of n !. (For example if the variables x, y, and z are
permuted in all 6 possible ways in the polynomial x + y - z then we get a total of 3
different polynomials: x + y − z, x + z - y, and y + z − x. Note that 3 is a factor of 6.) The
number of such polynomials is the index in the symmetric group Sn of the subgroup H of
permutations which preserve the polynomial. (For the example of x + y − z, the subgroup
H in S3 contains the identity and the transposition (xy).) So the size of H divides n !. With
the later development of abstract groups, this result of Lagrange on polynomials was
recognized to extend to the general theorem about finite groups which now bears his
name.
Lagrange did not prove his theorem; all he did, essentially, was to discuss some special
cases. The first complete proof of the theorem was provided by Abbati and published in
1803.
UNIT 8 7 Hours
Group codes
SYLLABUS
Group Codes: Decoding with Coset Leaders, Hamming Matrices
In computer science, group codes are a type of code. Group codes consist of n linear
block codes which are subgroups of Gn, where G is a finite Abelian group.
Construction
Group codes can be constructed by special generator matrices which resemble generator
matrices of linear block codes except that the elements of those matrices are
endomorphisms of the group instead of symbols from the code's alphabet. For example,
consider the generator matrix
The elements of this matrix are 2x2 matrices which are endomorphisms. In this scenario,
each codeword can be represented as where g1,...gr are the generators of
G.
In the field of coding theory, a coset leader is defined as a word of minimum weight in
any particular coset - that is, a word with the lowest amount of non-zero entries.
Sometimes there are several words of equal minimum weight in a coset, and in that case,
any one of those words may be chosen to be the coset leader.
Coset leaders are used in the construction of a standard array for a linear code, which can
then be used to decode received vectors. For a received vector y, the decoded message is
y - e, where e is the coset leader of y. Coset leaders can also be used to construct a fast
decoding strategy. For each coset leader u we calculate the syndrome uH′. When we
receive v we evaluate vH′ and find the matching syndrome. The corresponding coset
leader is the most likely error pattern and we assume that v+u was the codeword sent.
Example
1. The first row lists all codewords (with the 0 codeword on the extreme left)
2. Each row is a coset with the coset leader in the first column
3. The entry in the i-th row and j-th column is the sum of the i-th coset leader and the j-th
codeword.
For example, the [n,k]-code C3 = {0, 01101, 10110, 11011} has a standard array as
follows:
Note that the above is only one possibility for the standard array; had 00011 been chosen
as the first coset leader of weight two, another standard array representing the code would
have been constructed.
Note that the first row contains the 0 vector and the codewords of C3 (0 itself being a
codeword). Also, the leftmost column contains the vectors of minimum weight
enumerating vectors of weight 1 first and then using vectors of weight 2. Note also that
each possible vector in the vector space appears exactly once.
Because each possible vector can appear only once in a standard array some care must be
taken during construction. A standard array can be created as follows:
Hamming matrices
Hamming codes can be computed in linear algebra terms through matrices because
Hamming codes are linear codes. For the purposes of Hamming codes, two Hamming
matrices can be defined: the code generator matrix and the parity-check matrix
:
and
As mentioned above, rows 1, 2, & 4 of should look familiar as they map the data bits
to their parity bits:
The remaining rows (3, 5, 6, 7) map the data to their position in encoded form and there
is only 1 in that row so it is an identical copy. In fact, these four rows are linearly
independent and form the identity matrix (by design, not coincidence).
Also as mentioned above, the three rows of should be familiar. These rows are used to
compute the syndrome vector at the receiving end and if the syndrome vector is the null
vector (all zeros) then the received word is error-free; if non-zero then the value indicates
which bit has been flipped.
For the remainder of this section, the following 4 bits (shown as a column vector) will be
used as a running example:
Ring theory
Commutative rings are much better understood than noncommutative ones. Due to its
intimate connections with algebraic geometry and algebraic number theory, which
provide many natural examples of commutative rings, their theory, which is considered to
be part of commutative algebra and field theory rather than of general ring theory, is quite
different in flavour from the theory of their noncommutative counterparts. A fairly recent
trend, started in the 1980s with the development of noncommutative geometry and with
the discovery of quantum groups, attempts to turn the situation around and build the
theory of certain classes of noncommutative rings in a geometric fashion as if they were
rings of functions on (non-existent) 'noncommutative spaces'.
Elementary introduction
Definition
Formally, a ring is an Abelian group (R, +), together with a second binary operation *
such that for all a, b and c in R,
a * (b * c) = (a * b) * c
a * (b + c) = (a * b) + (a * c)
(a + b) * c = (a * c) + (b * c)
also, if there exists a multiplicative identity in the ring, that is, an element e such that for
all a in R,
a*e=e*a=a
then it is said to be a ring with unity. The number 1 is a common example of a unity.
The ring in which e is equal to the additive identity must have only one element. This
ring is called the trivial ring.
Rings that sit inside other rings are called subrings. Maps between rings which respect
the ring operations are called ring homomorphisms. Rings, together with ring
homomorphisms, form a category (the category of rings). Closely related is the notion of
ideals, certain subsets of rings which arise as kernels of homomorphisms and can serve to
define factor rings. Basic facts about ideals, homomorphisms and factor rings are
recorded in the isomorphism theorems and in the Chinese remainder theorem.
For example,
because 38 − 14 = 24, which is a multiple of 12. For positive n and non-negative a and b,
congruence of a and b can also be thought of as asserting that these two numbers have the
same remainder after dividing by the modulus n. So,
because both numbers, when divided by 12, have the same remainder (2). Equivalently,
the fractional parts of doing a full division of each of the numbers by 12 are the same:
0.1666... (38/12 = 3.1666..., 2/12 = 0.1666...). From the prior definition we also see that
their difference, a − b = 36, is a whole number (integer) multiple of 12 (n = 12, 36/12
= 3).
The properties that make this relation a congruence relation (respecting addition,
subtraction, and multiplication) are the following.
If
and
then:
Group axioms
Because a ≡ b (mod n) implies that gcd(a, n) = gcd(b, n), the notion of congruence
classes (mod n) which are relatively prime to n is well-defined.
Since gcd(a, n) = 1 and gcd(b, n) = 1 implies gcd(ab, n) = 1 the set of classes relatively
prime to n is closed under multiplication.
The natural mapping from the integers to the congruence classes (mod n) that takes an
integer to its congruence class (mod n) is a ring homomorphism. This implies that the
class containing 1 is the unique multiplicative identity, and also the associative and
commutative laws.
Notation
Structure
Powers of 2
Modulo 8 there are four relatively prime classes, 1, 3, 5 and 7. The square of each of
these is 1, so the Klein four-group.
Modulo 16 there are eight relatively prime classes 1, 3, 5, 7, 9, 11, 13 and 15.
is the 2-torsion subgroup (ie. the square of each element is 1),
so is not cyclic. The powers of 3, {1,3,9,11} are a subgroup of order 4, as are
the powers of 5, {1,5,9,13}. Thus
The pattern shown by 8 and 16 holds[1] for higher powers 2k, k > 2:
is the 2-torsion subgroup (so is not cyclic)
and the powers of 3 are a subgroup of order 2k − 2, so
Similarly, the group of units is the direct product of the groups corresponding
to each of the prime power factors:
Order
Exponent
The exponent is given by the Carmichael function λ(n), the least common multiple of the
orders of the cyclic groups. This means that if a and n are relatively prime,
Generators
Since all the n = 1, 2, ..., 7 are cyclic, another way to state this is: If n < 8
then has a primitive root. If n ≥ 8 has a primitive root unless n is
divisible by 4 or by two distinct odd primes.
In the general case there is one generator for each cyclic direct factor.
Table
This table shows the structure and generators of for small values of n. The
generators are not unique (mod n); e.g. (mod 16) both {−1, 3} and {−1, 5} will work. The
generators are listed in the same order as the direct factors.
For example take n = 20. means that the order of is 8 (i.e. there
are 8 numbers less than 20 and coprime to it); λ(20) = 4 that the fourth power of any
number relatively prime to 20 is ≡ 1 (mod 20); and as for the generators, 19 has order 2, 3
has order 4, and every member of is of the form 19a × 3b, where a is 0 or 1
and b is 0, 1, 2, or 3.
The powers of 19 are {±1} and the powers of 3 are {3, 9, 7, 1}. The latter and their
negatives (mod 20), {17, 11, 13, 19} are all the numbers less than 20 and prime to it. The
fact that the order of 19 is 2 and the order of 3 is 4 implies that the fourth power of every
member of is ≡ 1 (mod 20).