This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License
Vietnam Institute of Mathematics, Vietnam
In statistical theory, a statistic that is function of sample observations is used to estimate distribution parameter. This statistic is called unbiased estimate if its expectation is equal to theoretical parameter. Proving whether or not a statistic is unbiased estimate is very important but this proof may require a lot of efforts when statistic is complicated function. Therefore, this research facilitates this proof by proposing a theorem which states that the expectation of variable x > 0 is μ if and only if the limit of logarithm expectation of x approaches logarithm of μ. In order to make clear of this theorem, the research gives an example of proving correlation coefficient as unbiased estimate by taking advantages of this theorem.
Logarithm expectation, correlation coefficient, unbiased estimate.
 Douglas C. Montgomery, George C. Runger. Applied Statistics and Probability for Enginners, 3rd edition. Copyright 2003 © John Wiley & Son, Inc, 2003. ISBN: 0-471-20454-4.
 Jeff A. Bilmes. A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models. International Computer Science Institute, Berkeley CA, 94704 and Computer Science Division, Department of Electrical Engineering and Computer Science, U.C. Berkeley, TR-97-021, April 1998.
 Ronald E. Walpole, Raymond H. Myers, Sharon L. Myers, Keying Ye. Probability & Statistics for Engineers & Scientists, 9th edition. Copyright © 2012, 2007, 2002 Pearson Education, Inc. ISBN: 978-0-321-62911-1.