Hello,
I have a question regarding the second problem from the final exam from 2020. We are given n i.i.d. samples from a Bernoulli distribution with parameter , where we know that is between and for some and we are asked to accurately estimate the entropy of this distribution.
1) The proposed solution is to construct an estimate of the true parameter and then to use this for computing the entropy. This should be done by taking into account that . However, the proposed estimator is . I don't understand why it makes sense to have our estimator have a maximum value of . This means that if (eg: , this is perfectly fine according to our assumptions), our estimate will never be more than . I guess instead we'd want in order to make sure that our estimator is in the good range?
2) For the second part, I understand that is -subgaussian and the subsequent bound on (this is simply from Hoeffding's lemma applied for the Bernoulli random variable + subgaussian bound). Just to make sure, you obtain the bound on ? via the mean value theorem for ?
Thank you!
Andreea