Hi All,

When I run the logit command I get the following result

outcome | Coef. Std. Err. z P>|z| [95% Conf. Interval]
--------------------------------+----------------------------------------------------------------
indep_var | 3.147702 1.095392 2.87 0.004 1.000772 5.294632


When I run the same command with the or option I get the following output

outcome | Odds Ratio Std. Err. z P>|z| [95% Conf. Interval]
--------------------------------+----------------------------------------------------------------
indep_var| 23.2825 25.50347 2.87 0.004 2.720382 199.2642


I understand how the confidence interval is calculated from the log scale results - for example -

exp(3.147702 + 1.96*1.095392) approx.= exp(5.294632) = 199.2642


How does the numerical scale Std Err of 25.50347 relate to the log scale Std Err of 1.095392 and how would it be used to calculate the confidence interval on the numerical scale?

Thanks in advance,

Don