Talk:Generalized additive model
WikiProject Statistics  (Rated Startclass, Lowimportance)  


WikiProject Computer science  (Rated Startclass, Lowimportance)  


The following Wikipedia contributors may be personally or professionally connected to the subject of this article. Relevant policies and guidelines may include
conflict of interest,
autobiography, and
neutral point of view.

It is requested that a photograph be
included in this article to
improve its quality.
The Free Image Search Tool or Creative Commons Search may be able to locate suitable images on Flickr and other web sites. 
Potential Improvements
The area of GAMs is critically important in modern data mining, and this article could use a lot of additional work. I'll put in some more sections over the coming days to try to bring it more in line with the current state of the art. Also, because the GAM is a highly practical subject, I think it is worthwhile to discuss some practical matters related to this model. In particular, there is a lot of work out there related to using the GAM approach to perform functional decomposition of large datasets in order to discover the functional form of the phenomena that drive observed results. This is touched on in a few of the references, but is not discussed in the article itself, which is a shame.
Also, Tibschirani's original paper talks only about nonparametric methods, but semiparametric methods are also fairly common in recent years. The main reason for this is that semiparametric methods allow (often) for explanation of the causes behind these effects, and they are also much more able to control the complexity class of the model sought. — Preceding unsigned comment added by Vertigre ( talk • contribs) 20:01, 29 December 2015 (UTC)
Multiple regression vs GLM
Hmm, is the article in error? I've been taught that GAMs are extensions of Generalized Linear Models, not multiple regressions. Specifically, instead of the mean being a sum of component functions, it need only be related by a link function. (This construction contains the one given in the article.)  Fangz ( talk) 15:03, 15 May 2008 (UTC)
Further development of this article might be needed
I think this article only superficially touches this increasingly important area of applied statistics. In order to understand importance of GAM one should ask himself how realistic is that a particular variable has a linear effect, which is the restriction of GLM or linear models. And nonlinear least squares could be very timeconsuming and do not provide a clear inferential framework, not to mention convergence problems that could arise.
Off course overfitting as well as underfitting could be a problem, but there are many methods to adjust the model: CrossValidation (OCV), GCV, AIC, BIC or even through effect plots (especially in Rpackage mgcv). Simulations studies show that if GAM is used appropriately they are almost always outperform any other methods in vast variety of applications. Stats30 ( talk) 00:03, 9 March 2009 (UTC)
Link to Additive Models
This page links back to itself via a redirect. The link to Additive Models should be removed. —Preceding unsigned comment added by 131.181.251.66 ( talk) 04:07, 19 May 2009 (UTC)
Conflict of interest
@ SimonNWood: please note the policy on conflicts of interest. Tayste ( edits) 20:49, 20 July 2017 (UTC)
@ Tayste: can you be more specific? I've had a look but can't see my COI. SimonNWood ( talk) 21:03, 20 July 2017 (UTC)
Gaussian noise
Mostly it will sum to Gaussian noise except for specific inputs that incite some correlation in the function outputs. Then is it not some form of associative memory? As a simple example you have a locality sensitive hash whose output bit you view as +1,1. Weight each bit and sum to get a recalled value. To train, recall and calculate the error. Divide by the number of bits. Then add or subtract that as appropriate to each weight to make the error zero. Spreading out the error term that way decorrelates it when there is nonsimlar input, the error term fragments will sum to mean zero low level Gaussian noise. You can use predetermined random pattern of sign flipping applied the elements of a 1d vector followed by the fast Walsh Hadamard transform to get a random projection (RP.) Repeat for better quality. Then you can binarize the output of the RP to get a fast locality sensitive hash. Anyway if you understand these things you can see that associative memory=extreme learning machines=reservoir computing etc. — Preceding unsigned comment added by 113.190.221.54 ( talk) 11:33, 24 February 2019 (UTC)
 Does this have anything at all to do with generalized additive models?  Qwfp ( talk) 17:35, 24 February 2019 (UTC)
Generalized additive models with pairwise interactions (GA^{2}Ms)
Rich Caruana has been doing research on how GA^{2}Ms can be used as highly accurate and intelligible models for machine learning. See, for example:
 Accurate Intelligible Models with Pairwise Interactions (KDD '13)
 Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30day Readmission (KDD '15)
How can we incorporate this information into the article? Qzekrom (she/they • talk) 22:57, 7 November 2019 (UTC)