🎓How I Study AIHISA
📖Read
📄Papers📰Blogs🎬Courses
💡Learn
🛤️Paths📚Topics💡Concepts🎴Shorts
🎯Practice
🧩Problems🎯Prompts🧠Review
Search

Concepts2

Category

🔷All∑Math⚙️Algo🗂️DS📚Theory

Level

AllBeginnerIntermediateAdvanced
Filtering by:
#binary entropy
📚TheoryIntermediate

Shannon Entropy

Shannon entropy quantifies the average uncertainty or information content of a random variable in bits when using base-2 logarithms.

#shannon entropy#information gain#mutual information+12
📚TheoryIntermediate

Information Theory

Information theory quantifies uncertainty and information using measures like entropy, cross-entropy, KL divergence, and mutual information.

#entropy#cross-entropy#kl divergence+12