Talk:Generalized additive model

From Wikipedia
https://en.wikipedia.org/wiki/Talk:Generalized_additive_model
WikiProject Statistics (Rated Start-class, Low-importance)
WikiProject icon

This article is within the scope of the WikiProject Statistics, a collaborative effort to improve the coverage of statistics on Wikipedia. If you would like to participate, please visit the project page or join the discussion.

Start-Class article  Start  This article has been rated as Start-Class on the quality scale.
  Low  This article has been rated as Low-importance on the importance scale.
 
WikiProject Computer science (Rated Start-class, Low-importance)
WikiProject iconThis article is within the scope of WikiProject Computer science, a collaborative effort to improve the coverage of Computer science related articles on Wikipedia. If you would like to participate, please visit the project page, where you can join the discussion and see a list of open tasks.
Start-Class article  Start  This article has been rated as Start-Class on the project's quality scale.
  Low  This article has been rated as Low-importance on the project's importance scale.
 
Things you can help WikiProject Computer science with:

Potential Improvements

The area of GAMs is critically important in modern data mining, and this article could use a lot of additional work. I'll put in some more sections over the coming days to try to bring it more in line with the current state of the art. Also, because the GAM is a highly practical subject, I think it is worthwhile to discuss some practical matters related to this model. In particular, there is a lot of work out there related to using the GAM approach to perform functional decomposition of large datasets in order to discover the functional form of the phenomena that drive observed results. This is touched on in a few of the references, but is not discussed in the article itself, which is a shame.

Also, Tibschirani's original paper talks only about nonparametric methods, but semi-parametric methods are also fairly common in recent years. The main reason for this is that semi-parametric methods allow (often) for explanation of the causes behind these effects, and they are also much more able to control the complexity class of the model sought. — Preceding unsigned comment added by Vertigre ( talkcontribs) 20:01, 29 December 2015 (UTC)



Multiple regression vs GLM

Hmm, is the article in error? I've been taught that GAMs are extensions of Generalized Linear Models, not multiple regressions. Specifically, instead of the mean being a sum of component functions, it need only be related by a link function. (This construction contains the one given in the article.) -- Fangz ( talk) 15:03, 15 May 2008 (UTC)

Further development of this article might be needed

I think this article only superficially touches this increasingly important area of applied statistics. In order to understand importance of GAM one should ask himself how realistic is that a particular variable has a linear effect, which is the restriction of GLM or linear models. And non-linear least squares could be very time-consuming and do not provide a clear inferential framework, not to mention convergence problems that could arise.

Off course over-fitting as well as under-fitting could be a problem, but there are many methods to adjust the model: Cross-Validation (OCV), GCV, AIC, BIC or even through effect plots (especially in R-package mgcv). Simulations studies show that if GAM is used appropriately they are almost always outperform any other methods in vast variety of applications. Stats30 ( talk) 00:03, 9 March 2009 (UTC)

Link to Additive Models

This page links back to itself via a redirect. The link to Additive Models should be removed. —Preceding unsigned comment added by 131.181.251.66 ( talk) 04:07, 19 May 2009 (UTC)

Conflict of interest

@ SimonNWood: please note the policy on conflicts of interest. Tayste ( edits) 20:49, 20 July 2017 (UTC)

@ Tayste: can you be more specific? I've had a look but can't see my COI. SimonNWood ( talk) 21:03, 20 July 2017 (UTC)

Gaussian noise

Mostly it will sum to Gaussian noise except for specific inputs that incite some correlation in the function outputs. Then is it not some form of associative memory? As a simple example you have a locality sensitive hash whose output bit you view as +1,-1. Weight each bit and sum to get a recalled value. To train, recall and calculate the error. Divide by the number of bits. Then add or subtract that as appropriate to each weight to make the error zero. Spreading out the error term that way de-correlates it when there is non-simlar input, the error term fragments will sum to mean zero low level Gaussian noise. You can use predetermined random pattern of sign flipping applied the elements of a 1d vector followed by the fast Walsh Hadamard transform to get a random projection (RP.) Repeat for better quality. Then you can binarize the output of the RP to get a fast locality sensitive hash. Anyway if you understand these things you can see that associative memory=extreme learning machines=reservoir computing etc. — Preceding unsigned comment added by 113.190.221.54 ( talk) 11:33, 24 February 2019 (UTC)

Does this have anything at all to do with generalized additive models? -- Qwfp ( talk) 17:35, 24 February 2019 (UTC)

Generalized additive models with pairwise interactions (GA2Ms)

Rich Caruana has been doing research on how GA2Ms can be used as highly accurate and intelligible models for machine learning. See, for example:

How can we incorporate this information into the article? Qzekrom (she/they • talk) 22:57, 7 November 2019 (UTC)