I have some questions about multiple regression models in python:
Why is it necessary to apply a “dummy intercept” vector of ones to start for the Least Square method (OLS)? (I am refering to the use of X = sm.add_constant(X). I know that the Least square method is a system of derivatives equal to zero. Is it computed with some iterative method that make a “dummy intercept” necessary? Where can I find some informative material about the detail of the algorithm est = sm.OLS(y, X).fit()?
As far as I understood, scale.fit_transform produce a normalization of the data. Usually a normalization do not produce value higher than 1. Why, once scaled I see value that exceed 1?
Where is it possible to find a official documentation about python functions?
Thanks in advance
You can fit a line without constant term and you will get set of coefficients (dummy intercept is not necessary for iterative computation), but that might not be the best possible straight line which minimises the least square.