Abstract
This chapter is devoted to classical information theory fundamentals and its application to fading channels and channels with memory. The chapter starts with definitions of entropy, joint entropy, conditional entropy, relative entropy, mutual information, and channel capacity, followed by the information capacity theorem. We discuss the channel capacity of discrete memoryless channels, continuous channels, and channels with memory. Regarding the wireless channels, we describe how to calculate the channel capacity of flat fading and frequency-selective channels. We also discuss different optimum and suboptimum strategies to achieve channel capacity including the water-filling method, multiplexed coding and decoding, channel inversion, and truncated channel inversion. We also study different strategies for channel capacity calculation depending on what is known about the channel state information. Further, we explain how to model the channel with memory and describe McMillan–Khinchin model for channel capacity evaluation. After that, the fundamentals of linear blocks codes are introduced, followed by the binary LDPC coding fundamentals.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.