Entropy is a measure of how much information something provides
- Measured in bits
- One bit of entropy is amount of information it takes to encode a yes/no signal
Formula for Entropy
We need a function with the following properties:
- Continuous
- Should be 1 if all events are equally likely
- Should be 0 if outcome is certain
- Every time we double number of potential outcomes, entropy should increase by one
/ satisfies these properties.
If all events are equal probability, then:
Otherwise:
(note the minus sign)