What does information theory mean?
Definitions for information theory
in·for·ma·tion the·o·ry
This dictionary definitions page includes all the possible meanings, example usage and translations of the word information theory.
Princeton's WordNet
information theorynoun
(computer science) a statistical theory dealing with the limits and efficiency of information processing
GCIDE
information theorynoun
The science which studies the capacity of systems to contain, store, and transmit information{2 and 4}, and the factors such as noise and channel capacity that may affect the rate or accuracy of information transmission and reception.
Wiktionary
information theorynoun
A branch of applied mathematics and engineering involving the quantification of information sent over a communication channel, disregarding the meaning of the sent messages, exemplified by the noisy-channel coding theorem.
Wikipedia
Information theory
Information theory is the scientific study of the quantification, storage, and communication of information. The field was originally established by the works of Harry Nyquist and Ralph Hartley, in the 1920s, and Claude Shannon in the 1940s.: vii The field is at the intersection of probability theory, statistics, computer science, statistical mechanics, information engineering, and electrical engineering. A key measure in information theory is entropy. Entropy quantifies the amount of uncertainty involved in the value of a random variable or the outcome of a random process. For example, identifying the outcome of a fair coin flip (with two equally likely outcomes) provides less information (lower entropy) than specifying the outcome from a roll of a die (with six equally likely outcomes). Some other important measures in information theory are mutual information, channel capacity, error exponents, and relative entropy. Important sub-fields of information theory include source coding, algorithmic complexity theory, algorithmic information theory and information-theoretic security. Applications of fundamental topics of information theory include source coding/data compression (e.g. for ZIP files), and channel coding/error detection and correction (e.g. for DSL). Its impact has been crucial to the success of the Voyager missions to deep space, the invention of the compact disc, the feasibility of mobile phones and the development of the Internet. The theory has also found applications in other areas, including statistical inference, cryptography, neurobiology, perception, linguistics, the evolution and function of molecular codes (bioinformatics), thermal physics, molecular dynamics, quantum computing, black holes, information retrieval, intelligence gathering, plagiarism detection, pattern recognition, anomaly detection and even art creation.
ChatGPT
information theory
Information Theory is a branch of mathematics and computer science that focuses on quantifying, storing, and communicating information. It deals with the processes of encoding, transmission, decoding, and comprehension of messages or data. It seeks to understand the nature of information, how to handle it in practical situations like data compression and error detection, the limits and capacity of various communication channels, the measures of data complexity, and the volume or quantity of data. Introduced by Claude Shannon in 1948, information theory is fundamental for fields like telecommunication, data science, and computer engineering.
Wikidata
Information theory
Information theory is a branch of applied mathematics, electrical engineering, bioinformatics, and computer science involving the quantification of information. Information theory was developed by Claude E. Shannon to find fundamental limits on signal processing operations such as compressing data and on reliably storing and communicating data. Since its inception it has broadened to find applications in many other areas, including statistical inference, natural language processing, cryptography, neurobiology, the evolution and function of molecular codes, model selection in ecology, thermal physics, quantum computing, plagiarism detection and other forms of data analysis. A key measure of information is entropy, which is usually expressed by the average number of bits needed to store or communicate one symbol in a message. Entropy quantifies the uncertainty involved in predicting the value of a random variable. For example, specifying the outcome of a fair coin flip provides less information than specifying the outcome from a roll of a die. Applications of fundamental topics of information theory include lossless data compression, lossy data compression, and channel coding. The field is at the intersection of mathematics, statistics, computer science, physics, neurobiology, and electrical engineering. Its impact has been crucial to the success of the Voyager missions to deep space, the invention of the compact disc, the feasibility of mobile phones, the development of the Internet, the study of linguistics and of human perception, the understanding of black holes, and numerous other fields. Important sub-fields of information theory are source coding, channel coding, algorithmic complexity theory, algorithmic information theory, information-theoretic security, and measures of information.
U.S. National Library of Medicine
Information Theory
An interdisciplinary study dealing with the transmission of messages or signals, or the communication of information. Information theory does not directly deal with meaning or content, but with physical representations that have meaning or content. It overlaps considerably with communication theory and CYBERNETICS.
Matched Categories
Numerology
Chaldean Numerology
The numerical value of information theory in Chaldean Numerology is: 6
Pythagorean Numerology
The numerical value of information theory in Pythagorean Numerology is: 9
Examples of information theory in a Sentence
It’s not junk, it’s not damage that causes us to get old, we believe it’s a loss of information — a loss in the cell’s ability to read its original DNA so it forgets how to function — in much the same way an old computer may develop corrupted software. I call it the information theory of aging.
References
Translations for information theory
From our Multilingual Translation Dictionary
Get even more translations for information theory »
Translation
Find a translation for the information theory definition in other languages:
Select another language:
- - Select -
- 简体中文 (Chinese - Simplified)
- 繁體中文 (Chinese - Traditional)
- Español (Spanish)
- Esperanto (Esperanto)
- 日本語 (Japanese)
- Português (Portuguese)
- Deutsch (German)
- العربية (Arabic)
- Français (French)
- Русский (Russian)
- ಕನ್ನಡ (Kannada)
- 한국어 (Korean)
- עברית (Hebrew)
- Gaeilge (Irish)
- Українська (Ukrainian)
- اردو (Urdu)
- Magyar (Hungarian)
- मानक हिन्दी (Hindi)
- Indonesia (Indonesian)
- Italiano (Italian)
- தமிழ் (Tamil)
- Türkçe (Turkish)
- తెలుగు (Telugu)
- ภาษาไทย (Thai)
- Tiếng Việt (Vietnamese)
- Čeština (Czech)
- Polski (Polish)
- Bahasa Indonesia (Indonesian)
- Românește (Romanian)
- Nederlands (Dutch)
- Ελληνικά (Greek)
- Latinum (Latin)
- Svenska (Swedish)
- Dansk (Danish)
- Suomi (Finnish)
- فارسی (Persian)
- ייִדיש (Yiddish)
- հայերեն (Armenian)
- Norsk (Norwegian)
- English (English)
Word of the Day
Would you like us to send you a FREE new word definition delivered to your inbox daily?
Citation
Use the citation below to add this definition to your bibliography:
Style:MLAChicagoAPA
"information theory." Definitions.net. STANDS4 LLC, 2024. Web. 19 May 2024. <https://www.definitions.net/definition/information+theory>.
Discuss these information theory definitions with the community:
Report Comment
We're doing our best to make sure our content is useful, accurate and safe.
If by any chance you spot an inappropriate comment while navigating through our website please use this form to let us know, and we'll take care of it shortly.
Attachment
You need to be logged in to favorite.
Log In