Some Questions about Groups Coming from Information Theory
For each random n-vector there is an entropy vector of length 2^n-1. A fundamental question in information theory is to characterize the region formed by these entropic vectors. The region is bounded by Shannon's inequalities, but not tightly bounded for n>3. Chan and Yeung discovered that random vectors constructed from groups fill out the entropic region, so that information theoretic properties may be interpreted to give properties of groups and combinatorial results for groups may be used to better understand the entropic region. I will elaborate on these connections and present some simple and interesting questions about groups that arise.