The processing of information via the use of applied mathematics and electrical engineering
- Audio coding - the translation of auditory information into digital code
- Channel coding - code used to protect information over a channel by correcting errors resulting from noise or other interference
- Codes - rules for converting one piece of information into another
- Communication channels - a physical or logical connection between two points that allows for the exchange of an information signal
- Decoding - translating from an coded message into the original language or form
- Encoding - the process by which information from a source is changed into symbols to be communicated
- Error compensation - the encoding or transmission of extra information or code to compensate for possible errors
- Genetic communication - communication of information in a form analagous to biological genes, such as in a sequential or additive fashion
- Hamming distance - the number of positions in two strings of equal length which have different corresponding symbols
- Hamming weight - the number of symbols in a string that are different from the zero symbol in the utilized alphabet
- Information entropy - the level of uncertainty associated with a random variable (often refers to the "Shannon entropy")
- Mutual information - occasionally called transinformation, the quantity that measures the mutual dependence of two random variables
- Rate distortion theory - the branch of information theory which explains lossy data compression and which determines the minimal amount of entropy that should be communicated over a channel
- Rate-distortion - the number of bits per data sample to be stored or transmitted and the amount of distortion in that sample
- Speech coding - the use of the data compression of digital audio signals to encode speech
This category has the following 16 subcategories, out of 16 total.
Pages in category "Information theory"
The following 65 pages are in this category, out of 65 total.