Skip to content

Entropy is a measure of the uncertainty in a random variable. This application calculates the entropy of text. The current example calculates the entropy of sequence "TTTAAGCC". In the context of information theory the term "Entropy" refers to the Shannon entropy.

License

Notifications You must be signed in to change notification settings

Gagniuc/Entropy-of-Text

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

30 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Entropy of text

Entropy is a measure of the uncertainty in a random variable. This application calculates the entropy of text. The current example calculates the entropy of sequence "TTTAAGCC". In the context of information theory the term "Entropy" refers to the Shannon entropy:

Entropy

Which can also be written as:

Entropy

Where n represents the total number of symbols in the alphabet of a sequence and pi represents the probability of occurrence of a symbol i found in the alphabet. For more detailed information on entropy please see the specialized chapter from the book entitled Algorithms in Bioinformatics: Theory and Implementation.

Live demo: https://gagniuc.github.io/Entropy-of-Text/

References

  • Paul A. Gagniuc. Algorithms in Bioinformatics: Theory and Implementation. John Wiley & Sons, Hoboken, NJ, USA, 2021, ISBN: 9781119697961.

About

Entropy is a measure of the uncertainty in a random variable. This application calculates the entropy of text. The current example calculates the entropy of sequence "TTTAAGCC". In the context of information theory the term "Entropy" refers to the Shannon entropy.

Topics

Resources

License

Stars

Watchers

Forks

Sponsor this project

 

Packages

No packages published

Languages