Mathematical Foundations of Information Theory

Mathematical Foundations of Information Theory

A. I. Khinchin
你有多喜歡這本書?
文件的質量如何?
下載本書進行質量評估
下載文件的質量如何?
The first comprehensive introduction to information theory, this book places the work begun by Shannon and continued by McMillan, Feinstein, and Khinchin on a rigorous mathematical basis. For the first time, mathematicians, statisticians, physicists, cyberneticists, and communications engineers are offered a lucid, comprehensive introduction to this rapidly growing field.
In his first paper, Dr. Khinchin develops the concept of entropy in probability theory as a measure of uncertainty of a finite “scheme,” and discusses a simple application to coding theory. The second paper investigates the restrictions previously placed on the study of sources, channels, and codes and attempts “to give a complete, detailed proof of both … Shannon theorems, assuming any ergodic source and any stationary channel with a finite memory.”
Partial Contents: I. The Entropy Concept in Probability Theory — Entropy of Finite Schemes. The Uniqueness Theorem. Entropy of Markov chains. Application to Coding Theory. II. On the Fundamental Theorems of Information Theory — Two generalizations of Shannon’s inequality. Three inequalities of Feinstein. Concept of a source. Stationarity. Entropy. Ergodic sources. The E property. The martingale concept. Noise. Anticipation and memory. Connection of the channel to the source. Feinstein’s Fundamental Lemma. Coding. The first Shannon theorem. The second Shannon theorem.
年:
1957
版本:
1st Dover
出版商:
Dover Publications
語言:
english
頁數:
128
ISBN 10:
0486604349
ISBN 13:
9780486604343
系列:
Dover Books on Mathematics
文件:
PDF, 1.86 MB
IPFS:
CID , CID Blake2b
english, 1957
線上閱讀
轉換進行中
轉換為 失敗

最常見的術語