TY - BOOK ID - 4863494 TI - Transmitting and Gaining Data : Rudolf Ahlswede’s Lectures on Information Theory 2 AU - Ahlswede, Rudolf. AU - Ahlswede, Alexander. AU - Althöfer, Ingo. AU - Deppe, Christian. AU - Tamm, Ulrich. PY - 2015 SN - 9783319125237 3319125222 9783319125220 3319125230 PB - Cham : Springer International Publishing : Imprint: Springer, DB - UniCat KW - Mathematics. KW - Information and Communication, Circuits. KW - Mathématiques KW - Engineering & Applied Sciences KW - Mathematics KW - Physical Sciences & Mathematics KW - Computer Science KW - Mathematical Theory KW - Information theory. KW - Artificial intelligence. KW - AI (Artificial intelligence) KW - Artificial thinking KW - Electronic brains KW - Intellectronics KW - Intelligence, Artificial KW - Intelligent machines KW - Machine intelligence KW - Thinking, Artificial KW - Communication theory KW - Communication KW - Cybernetics KW - Bionics KW - Cognitive science KW - Digital computer simulation KW - Electronic data processing KW - Logic machines KW - Machine theory KW - Self-organizing systems KW - Simulation methods KW - Fifth generation computers KW - Neural computers KW - Math KW - Science UR - https://www.unicat.be/uniCat?func=search&query=sysid:4863494 AB - The calculation of channel capacities was one of Rudolf Ahlswede's specialties and is the main topic of this second volume of his Lectures on Information Theory. Here we find a detailed account of some very classical material from the early days of Information Theory, including developments from the USA, Russia, Hungary and which Ahlswede was probably in a unique position to describe the German school centered around his supervisor Konrad Jacobs. These lectures made an approach to a rigorous justification of the foundations of Information Theory. This is the second of several volumes documenting Rudolf Ahlswede's lectures on Information Theory. Each volume includes comments from an invited well-known expert. In the supplement to the present volume, Gerhard Kramer contributes his insights. Classical information processing concerns the main tasks of gaining knowledge and the storage, transmission and hiding of data. The first task is the prime goal of Statistics. For transmission and hiding of data, Shannon developed an impressive mathematical theory called Information Theory, which he based on probabilistic models. The theory largely involves the concept of codes with small error probabilities in spite of noise in the transmission, which is modeled by channels. The lectures presented in this work are suitable for graduate students in Mathematics, and also for those working in Theoretical Computer Science, Physics, and Electrical Engineering with a background in basic Mathematics. The lectures can be used as the basis for courses or to supplement courses in many ways. Ph.D. students will also find research problems, often with conjectures, that offer potential subjects for a thesis. More advanced researchers may find questions which form the basis of entire research programs. ER -