Information Theory

From LNTwww
Revision as of 13:10, 26 September 2021 by Guenter (talk | contribs)

Since the early beginnings of communications as an engineering discipline, many engineers and mathematicians have sought to find a quantitative measure of

  • the $\rm Information$  (in general: "the knowledge of something") contained in a  $\rm message$  (here we understand  "a collection of symbols and/or states").


The  (abstract)  information is communicated by the  (concrete)  message and can be seen as an interpretation of a message.

Claude Elwood Shannon  succeeded in 1948 in establishing a consistent theory of the information content of messages,  which was revolutionary in its time and created a new, still highly topical field of science:  the theory named after him  $\text{Shannon's Information Theory}$.

The subject matter corresponds to a  $\text{lecture with two semester hours per week (SWS) and one additional SWS exercise}$.

Here is a table of contents based on the  $\text{four main chapters}$  with a total of  $\text{13 individual chapters}$.


Contents

In addition to these theory pages, we also offer exercises and multimedia modules that could help to clarify the teaching material:



$\text{More links:}$



$(1)$    $\text{Bibliography to the book}$

$(2)$    $\text{Notes on the authors and materials used in the preparation of the book}$



$(2)$    $\text{General notes about the book}$   (Authors,  other participants,  materials as a starting point for the book,  list of sources)