To answer your question directly, the accepted answer in the thread you reference, does indeed implicitly make the assumption of normally distributed errors:
\begin{array}{l}
\mathbf{y} = \mathbf{X} \mathbf{\beta} + \mathbf{\epsilon} \\
\mathbf{\epsilon} \sim N(0, \sigma^2 \mathbf{I}),
\end{array}
This is not necessary. For example, note, that the Gauss-Markov theorem only assumes that the $\epsilon$'s:
- Have zero mean: $\mathbf{E}(\epsilon_{i}) = 0$
- Have constant variance: $\text{Var}(\epsilon_{i}) = \sigma^{2} < \infty $
- Are uncorrelated with each other
Yet, proof of the theorem involves derivation of the variance of $\hat{\beta}$. Standard error could be defined from there.
The assumption of normality of the errors makes the least squares solution also the maximum likelihood estimate. From there, the distributions of the parameters can be derived and confidence intervals defined. See this nice explanation or one from this site. Furthermore, since for the least squares case, we arrive at a $t$ pivot, the intervals are “exact” while for the generalized linear model, we’d rely on the asymptotic covariance matrix of $\hat{\beta}$.