There are two forms of Logistic Regression used in literature. In this post, I will build a bridge between these two forms and show they are equivalent.
Logistic Function & Logistic Regression
The common definition of Logistic Function is as follows:

![Rendered by QuickLaTeX.com P(x) \in [0,1]](https://www.hongliangjie.com/wp-content/ql-cache/quicklatex.com-88f94b8f82e0c5825bb11baec49d3c15_l3.png)






The Equivalence of Two Forms of Logistic Regression
At first glance, the form (3) and the form (4) looks very different. However, the equivalence between these two forms can be easily established. Starting from the form (3), we can have:


Logistic Loss
Since we establish the equivalence of two forms of Logistic Regression, it is convenient to use the second form as it can be explained by a general classification framework. Here, we assume is the label of data and
is a feature vector. The classification framework can be formalized as follows:



References
[1] Trevor Hastie, Robert Tibshirani, and Jerome Friedman. The Elements of Statistical Learning. Springer Series in Statistics. Springer New York Inc., New York, NY, USA, 2001.
[2] Tom M. Mitchell. Machine learning. McGraw Hill series in computer science. McGraw-Hill, 1997.
[3] Jason D. M. Rennie. Logistic Regression. http://people.csail.mit.edu/jrennie/writing, April 2003.