Difference between revisions of "Information theory"
Jump to navigation
Jump to search
m (→Contents) |
|||
Line 20: | Line 20: | ||
<math>I = H_{ref} - H_{obs}</math> | <math>I = H_{ref} - H_{obs}</math> | ||
+ | |||
+ | * {{WP|Information_theory}} | ||
+ | * {{WP|Entropy_in_thermodynamics_and_information_theory}} | ||
+ | * {{WP|Mutual_information}} | ||
Revision as of 14:36, 28 October 2012
Information theory
This page is a placeholder, or under current development; it is here principally to establish the logical framework of the site. The material on this page is correct, but incomplete.
This is an introduction to information theory for the bioinformatics lab.
Contents
Failed to parse (MathML with SVG or PNG fallback (recommended for modern browsers and accessibility tools): Invalid response ("Math extension cannot connect to Restbase.") from server "https://wikimedia.org/api/rest_v1/":): {\displaystyle H = - \sum_{i=0}^n p_i \log_{2} p_i}
Failed to parse (MathML with SVG or PNG fallback (recommended for modern browsers and accessibility tools): Invalid response ("Math extension cannot connect to Restbase.") from server "https://wikimedia.org/api/rest_v1/":): {\displaystyle I = H_{ref} - H_{obs}}
Further reading and resources
Wang & Samudrala (2006) Incorporating background frequency improves entropy-based residue conservation measures. BMC Bioinformatics 7:385. (pmid: 16916457) |
Dou et al. (2010) Several appropriate background distributions for entropy-based protein sequence conservation measures. J Theor Biol 262:317-22. (pmid: 19808039) |
Shannon's "Mathematical Theory of Communication" (at Bell labs)