Recall from lectures that information theory underpins a vast number of concepts in computer and information systems. Concepts such as data transmission, compression, estimating effort, and machine learning, heavily lean on information theory concepts.
In this lab, you will be working on the following exercises:
- computing the entropy of some small systems,
- creating a Huffman code for these systems (and examining their efficiency),
- and finally using a Huffman coding to encode and decode data.