Add hessian/fisher matrix for TweedieRegression
See original GitHub issueInternal feature
A prerequisite for several possible additional (Newton-like) solvers for TweedieRegression
is the hessian or fisher matrix.
Issues/PR that depend on this
Discussion
Hessian vs Fisher
The Fisher information matrix is the expected hessian. It is always (semi-) positive definite (as it is the variance of the score). This has advantages for stability of solvers.
Full matrix vs diagonal part
The hessian and fisher matrix (without penalty) for GLMs has the following special structure X.T @ W @ X
where W
is a diagonal matrix (depending on the family, the link function and the coefficients/weights). Some solvers might prefer to get the whole matrix, others only the diagonal part W
.
Issue Analytics
- State:
- Created 4 years ago
- Reactions:1
- Comments:8 (8 by maintainers)
Top Results From Across the Web
Basic question about Fisher Information matrix and ...
Regarding your main question: No, it's not correct that the observed Fisher information can be found by inverting the (negative) Hessian.
Read more >Making Tweedie's compound Poisson model more accessible
1, below, for an example. We conclude by calculating Fisher's information matrix for (\varvec{\alpha },\gamma ) in our gamma GLM. We ...
Read more >ptmixed: Poisson-Tweedie Generalized Linear Mixed Model
Default is 5. hessian. Logical value. If TRUE, the hessian matrix is evaluated at the MLE to derive the observed Fisher information matrix....
Read more >A Connection between the Fisher and Hessian
In practice, we have found that the Fisher and Hessian return similar examples when used to compute Itest,i.
Read more >The HPGENSELECT Procedure - SAS Support
It can be characterized in terms of the distribution mean parameter , dispersion parameter , and power parameter p. For more information about...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Not necessarily pushing for the addition of additional solvers aside for LBFGS, but having multiple solvers is useful to check that the implementation is correct. Here we can check it indirectly with Ridge and ElasticNet so it’s indeed less of a concern.
For irls and newton_cg certainly. coordinate descent solver is still necessary if we want L1 penalty support irrespective of it’s performance compared to LBFGS…