Download (a) (b) - Brandeis

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Non-coding DNA wikipedia , lookup

Promoter (genetics) wikipedia , lookup

Ancestral sequence reconstruction wikipedia , lookup

Point mutation wikipedia , lookup

Bisulfite sequencing wikipedia , lookup

Artificial gene synthesis wikipedia , lookup

Molecular evolution wikipedia , lookup

Homology modeling wikipedia , lookup

Transcript
More about Markov model.
The simple two-state model from the previous lecture was defined by
four transition probabilities (of which two were independent). They can
be described by the so called “transition matrix”
P=
A
B
A
1-p
p
B
q
1-q
Its equilibrium state could be defined by the “detailed equilibrium
condition”:
number of customers leaving the state A at the step k, (fk •p). equals
the number of customers joining the state A by leaving state B:
f • p =(1-f) • q 
f= q/(p+q)
1
In general, with s possible states, we can write the transition probability
matrix as
 p11
p
P   21
 :

 ps1
p12
p13
p22
:
p23
:
ps 2
ps 3
...... p1s 
...... p2s 

:
: 

...... pss 
Any row in this matrix corresponds to the state from which transition is made, and
any column corresponds to the state to which transition is made. Thus probabilities
in any particular row should sum to 1, while there is no such requirement for the
columns (Why? Explain in plain words) . Together with an initial probability
distribution, transition matrix determines the Markov process.
2
Example:Random walks (Evens&Grant, p136: “Random walk theory
underpins BLAST theory”
The “practical” example involves a gambler ( G) having i dollars and its
adversary (A) with (s-i) dollars. At each step a coin with probability p for
heads is tossed. If heads come up, A gives G one dollar, and if tails up,
then G gives a dollar to A. This continues until the random variable
reaches either 0 or s.
The random variable is a current fortune of the gambler taking values
0,1,2,…,s dollars.
G:
0 1
1 0
q 0

0 q

P  : :
0 0

0 0
 0 0
2 3
0
...
0
0
p 0 ... 0
0 p ... 0
: : .:. :
0
... q
0
0
:
0
0
0
0
s-1 s $
...
...
0 q
0 0
0
0 0

0 0

: :
p 0

0 p
0 1 
0
3
Example:Random walks (Evens&Grant, p136: “Random walk theory
underpins BLAST theory”
If G is left without money, he stays in this state with
p00=1, and all other transitions become impossible
1 0
q 0

0 q

P  : :
0 0

0 0
 0 0
0
...
0
0
p 0 ... 0
0 p ... 0
: : .:. :
0
... q
0
0
:
0
0
0
0
If G has $1, he throws a coin, and will either gain $1
(probability p) or will lose it (q). He will never stay with $1,
and thus p11=0
...
...
0 q
0 0
0
0 0

0 0

: :
p 0

0 p
0 1 
0
0 dollars
1 dollar
2 dollars
s-1 dollar
s dollars
4
The transition matrix are widely used in bioinformatics, for sequences analysis. Make
sure that you understand why P for the “gambler” has exactly this shape. Why, for
instance, p3,5=0?
What is the probability of moving from state i to the state j in two steps?
We have to sum up the probabilities for all possible “trajectories”
connecting i and j. The result is defined by the product P*P of two matrixes
(not two numbers!)
pij   pim pmj
(1)
m
{m}
i
The product of two matrixes, a and b is the
matrix c whose components are
j
cij   aim bmj
m
With this in mind, Eq. 1 can be written as a product
Similarly, the n-step transition probability is
P ( 2)  P  P  P 2
P(n)  P n
5
Graphical representation of the Markov Chain
0
1
2
3
1/3
0
1
2
3
1/3
6
Graphical representation
The corrected diagram
0
0
1
1
2
2
3
3
1/3
1/3
1/6
1/6
1/6
1/3
1/3
0
1
2 3
0
1
2 3
1/6
1/6
7
Working in groups:
• Find the product of matrixes A and B
1 3 2 
A  0 2 4 


5 0 0 
and
0 1 0 
B  0 2 4 


5 3 6 
• open Lect7/Math/Lect7_MatrixExample and use it for practice
• Draw a graphical representation for the transition matrixes between the states E1, E2 , E3
(a) and E4 (for b)
(a)
.2 .1 .7 
.5 .3 .2 


.6 0 .4 
(b)
.2 0
.5 .3

.6 0

0 0
.8
0
.1
0
0
.2 

.3 

1
8
One application of Markov Model for sequence analysis
In the human genome wherever the dinucleotide CG occurs (frequently
written CpG to distinguish it from the C-G base pair across the two
strands) the C nucleotide (cytosine) is typically chemically modified.
There is a relatively high chance of this modification that mutates C into
a T, and in general CpG dinucleotides are rarer in the genome than
would be expected from the independent probabilities of C and G*. For
biologically important reasons the mutation modification process is
suppressed in short stretches of the genome, such as around the
promoters or ‘start’ regions of many genes. In these regions we see
many more CpG dinucleotides than elsewhere. Such regions are called
CpG islands.
You can find the tool for searching the CpG islands at:
http://bioinformatics.org/sms/cpg_island.html
*The CpG dinucleotide is present at approximately 20% of its expected frequency in vertebrate* genomes, a
deficiency thought due to a high mutation rate from the methylated form of CpG to TpG and CpA.
9
What sort of probabilistic model might we use for
CpG island regions? It should be a model that
generates sequences in which the probability of a
symbol depends on the previous symbol. The
simplest such model is a classical Markov chain. A
Markov chain for DNA can be drawn like this:
10
where there is a state for each of the four letters A, C, G, and
T in the DNA alphabet. The transition probabilities pst are
associated with each arrow in the figure and determine the
probabilities of a certain residue following another residue (or
one “state” following another “state”). Markov models for the
CpG island are illustrated below. From a set of human DNA
sequences a total of 48 putative CpG islands have been
extracted. Two Markov chain models have been developed,
one for the regions labeled as CpG islands (the ‘+’ model)
and the other from the remainder of the sequence (the ‘-’
model). The transition probabilities for each model were set
using the equation

c
pst   st
 cst '
t'
11
where cts is the number of times letter t followed letter s in the island
regions. These are the maximum likelihood estimators for the transition
probabilities. The resulting tables are
12
13
The equation (2) therefore becomes
P( x )  P( xL | xL 1 )P( xL 1 | xL  2 )    P( x2 | x1 )
(3)
Eq. (3) is used to calculate the likelihood of the sequence. For instance, we use
(3) to obtain the likelihood values of a sequence under “+” model and “-”
model respectively. Comparing them, we can conclude if the sequence belongs
to the CpG island region or not. More specifically, we can calculate the logodds ratio:
p( xi|xi1)
l
P
(
x
|
model

)
S ( x)  log
  log 
(4)
(
x
|
x
)
p
P( x | model ) i1
i i1
14
Exercise 1
15
A brief introduction to the HMM

Very efficient programs for searching a text for a combination of words
are widely available The same methods can be used for searching for patterns in
biological sequences, but often they fail. Why?

Biological ‘spelling’ is much more sloppy than English spelling: proteins
with the same function from two different organisms are most likely spelled
differently, that is, the two amino acid sequences differ. It is not rare that two
such homologous sequences have less than 30% identical amino acids. Similarly
in DNA many interesting signals vary greatly even within the same genome.
Some well-known examples are ribosome binding sites and splice sites, but the
list is long.

Fortunately there are usually still some subtle similarities between two
such sequences, and the question is how to detect these similarities.
16
• The variation in a family of sequences can be described statistically, and
this is the basis for most methods used in biological sequence analysis
•A hidden Markov model (HMM) is a statistical model, which is very
well suited for many tasks in molecular biology, although they have been
mostly developed for speech recognition since the early 1970s.
•The most popular use of the HMM in molecular biology is as a
‘probabilistic profile’ of a protein family, which is called a profile HMM.
From a family of proteins (or DNA) a profile HMM can be made for
searching a database for other members of the family.
•Probably the main contribution of HMM comparative to some other
methods is that the profile HMM treats gaps in a systematic way.
17
From Regular expressions to HMM
In programs like grep, JavaScript and Perl, regular expressions can be used
for searching text files for a pattern.
Using regular expressions is a very elegant and efficient way to search for
some protein families, but difficult for other. The difficulties arise because protein
spelling is much more free than English spelling. Therefore the regular
expressions sometimes need to be very broad and complex. Imagine a DNA motif
like this:
A regular expression for this is
[AT] [CG] [AC] [ACGT]* A [TG] [GC] , meaning that the first position is A or
T, the second C or G, and so forth. The term ‘[ACGT]*’ means that any of the
four letters can occur any number of times.
18
The problem with the above regular expression is that it does not in any way
distinguish between the highly implausible sequence
“TGCT- - AGG” and the consensus sequence “ACAC- -ATC”.
It is possible to make the regular expression more discriminative by splitting it
into several different ones, but it easily becomes messy. The alternative is to
score sequences by how well they fit the alignment.
19
Such a scoring is shown in the diagram below.
This is the HMM derived from the same alignment
Transition Probabilities
States
‘insertion’ state is represented by the state above the other
states.
20
It is now easy to score the consensus sequence ACACATC. The probabilities of
different states (bases) must be multiplied by the transition probabilities:
P ( ACACATC )  0.8  1  0.8  1  0.8  0.6  0.4  0.6  1  1  0.8  1  0.8
 4.7102
Making the same calculation for the exceptional sequence yields only 0.0023 10-2
21
Probabilities and log-odds scores for the 5 sequences in the alignment
and for the consensus sequence and the ‘exceptional’ sequence:
22
What is the Log-odds score?
The probability itself is not the most convenient number to use as a score, and the logodds score shown in the last column of the table is usually better. It is the logarithm
of the probability of the sequence divided by the probability according to a null
model. The null model is one that treats the sequences as random strings of
nucleotides, so the probability of a sequence of length L is
23
The probabilities of the model in the previous picture have been turned
into log-odds by taking the logarithm of each nucleotide probability and
subtracting log 0.25. The transition probabilities have been converted to
simple logs.
When a sequence fits the motif very well the log-odds is high. When it’s
very unlikely, the log-odds score becomes negative.
Here is an example for the consensus sequence:
24
I will stop right here. This discussion was based on a very
well-written article (see the “resources” and a link on my
site),
Reading assignment: pp.8-13 of that article, “Profile HMM” section. This reading
will prepare you for better understanding of the project presentation devoted to
HMM
25