• No results found

Paper: Regression Analysis III Module: Components of GLM

N/A
N/A
Protected

Academic year: 2022

Share "Paper: Regression Analysis III Module: Components of GLM"

Copied!
53
0
0

Loading.... (view fulltext now)

Full text

(1)

Subject: Statistics

Paper: Regression Analysis III Module: Components of GLM

Regression Analysis III 1 / 14

(2)

Development Team

Principal investigator: Dr. Bhaswati Ganguli,Professor, Department of Statistics, University of Calcutta

Paper co-ordinator: Dr. Bhaswati Ganguli,Professor, Department of Statistics, University of Calcutta

Content writer: Sayantee Jana, Graduate student, Department of Mathematics and Statistics, McMaster University Sujit Kumar Ray,Analytics professional, Kolkata

Content reviewer: Department of Statistics, University of Calcutta

(3)

Components of a GLM

There are 3 components of a GLM :

I Random component

I Systematic component

I Link function

Regression Analysis III 3 / 14

(4)

Components of a GLM

There are 3 components of a GLM :

I Random component

I Systematic component

I Link function

(5)

Components of a GLM

There are 3 components of a GLM :

I Random component

I Systematic component

I Link function

Regression Analysis III 3 / 14

(6)

Components of a GLM

There are 3 components of a GLM :

I Random component

I Systematic component

I Link function

(7)

The random component

I ConsiderY to be the response variable.

I Y = (Y1, Y2, . . . , Yn) : Yi’s are independently distributed.

I We are considering one parameter exponential families here so we assume the distribution of Y to belong to that family and have the form :

f(y) =e(

yθ−b(θ)

a(φ) +C(y,φ))

I θ : natural or canonical parameters

I φ: dispersion parameters

Regression Analysis III 4 / 14

(8)

The random component

I ConsiderY to be the response variable.

I Y = (Y1, Y2, . . . , Yn) : Yi’s are independently distributed.

I We are considering one parameter exponential families here so we assume the distribution of Y to belong to that family and have the form :

f(y) =e(

yθ−b(θ)

a(φ) +C(y,φ))

I θ : natural or canonical parameters

I φ: dispersion parameters

(9)

The random component

I ConsiderY to be the response variable.

I Y = (Y1, Y2, . . . , Yn) : Yi’s are independently distributed.

I We are considering one parameter exponential families here so we assume the distribution of Y to belong to that family and have the form :

f(y) =e(

yθ−b(θ)

a(φ) +C(y,φ))

I θ : natural or canonical parameters

I φ: dispersion parameters

Regression Analysis III 4 / 14

(10)

The random component

I ConsiderY to be the response variable.

I Y = (Y1, Y2, . . . , Yn) : Yi’s are independently distributed.

I We are considering one parameter exponential families here so we assume the distribution of Y to belong to that family and have the form :

f(y) =e(

yθ−b(θ)

a(φ) +C(y,φ))

I θ : natural or canonical parameters

I φ: dispersion parameters

(11)

Examples of One parameter Exponential family

I Normal: f(y) =e(

yµ−µ2/2 σ2 +[−y2

2−log 2πσ])

I Binomial: f(y) =e

(ylog1−ππ +nlog(1−π)+log

n y

)

Regression Analysis III 5 / 14

(12)

Examples of One parameter Exponential family

I Normal: f(y) =e(

yµ−µ2/2 σ2 +[−y2

2−log 2πσ])

I Binomial: f(y) =e

(ylog1−ππ +nlog(1−π)+log

n y

)

(13)

Members of the one parameter exponential family

Other members of the exponential family

I Poisson

I Binomial

I Negative Binomial

I Gamma

I Inverse Gaussian

Regression Analysis III 6 / 14

(14)

Members of the one parameter exponential family

Other members of the exponential family

I Poisson

I Binomial

I Negative Binomial

I Gamma

I Inverse Gaussian

(15)

Members of the one parameter exponential family

Other members of the exponential family

I Poisson

I Binomial

I Negative Binomial

I Gamma

I Inverse Gaussian

Regression Analysis III 6 / 14

(16)

Members of the one parameter exponential family

Other members of the exponential family

I Poisson

I Binomial

I Negative Binomial

I Gamma

I Inverse Gaussian

(17)

Members of the one parameter exponential family

Other members of the exponential family

I Poisson

I Binomial

I Negative Binomial

I Gamma

I Inverse Gaussian

Regression Analysis III 6 / 14

(18)

Systematic component

I Consists of the explanatory variable x and a linear function of x andβ

I x= (x1, x2, . . . , xp) : set ofp explanatory variables

I β = (β1, β2, , βp) : parameter vector

I η =x0β =Pp

j=1βjxj : linear predictor

(19)

Systematic component

I Consists of the explanatory variable x and a linear function of x andβ

I x= (x1, x2, . . . , xp) : set ofp explanatory variables

I β = (β1, β2, , βp) : parameter vector

I η =x0β =Pp

j=1βjxj : linear predictor

Regression Analysis III 7 / 14

(20)

Systematic component

I Consists of the explanatory variable x and a linear function of x andβ

I x= (x1, x2, . . . , xp) : set ofp explanatory variables

I β = (β1, β2, , βp) : parameter vector

I η =x0β =Pp

j=1βjxj : linear predictor

(21)

Systematic component

I Consists of the explanatory variable x and a linear function of x andβ

I x= (x1, x2, . . . , xp) : set ofp explanatory variables

I β = (β1, β2, , βp) : parameter vector

I η =x0β =Pp

j=1βjxj : linear predictor

Regression Analysis III 7 / 14

(22)

Link function

I It is a link between the random and systematic components.

I It is a function that specifies the relationship between the expected value of the random component and the systematic component.

I ηi =g(µi) whereµi=E(Yi)

and, g(.) is a continuous, monotone and differentiable function

I Identity link : g(µi) =µi

I Canonical link : g(µ) =θ

(23)

Link function

I It is a link between the random and systematic components.

I It is a function that specifies the relationship between the expected value of the random component and the systematic component.

I ηi =g(µi) whereµi=E(Yi)

and, g(.) is a continuous, monotone and differentiable function

I Identity link : g(µi) =µi

I Canonical link : g(µi) =θi

Regression Analysis III 8 / 14

(24)

Link function

I It is a link between the random and systematic components.

I It is a function that specifies the relationship between the expected value of the random component and the systematic component.

I ηi =g(µi) whereµi=E(Yi)

and, g(.) is a continuous, monotone and differentiable function

I Identity link : g(µi) =µi

I Canonical link : g(µ) =θ

(25)

Link function

I It is a link between the random and systematic components.

I It is a function that specifies the relationship between the expected value of the random component and the systematic component.

I ηi =g(µi) whereµi=E(Yi)

and, g(.) is a continuous, monotone and differentiable function

I Identity link : g(µi) =µi

I Canonical link : g(µi) =θi

Regression Analysis III 8 / 14

(26)

Link function

I It is a link between the random and systematic components.

I It is a function that specifies the relationship between the expected value of the random component and the systematic component.

I ηi =g(µi) whereµi=E(Yi)

and, g(.) is a continuous, monotone and differentiable function

I Identity link : g(µi) =µi

I Canonical link : g(µ) =θ

(27)

Link function

I It is a link between the random and systematic components.

I It is a function that specifies the relationship between the expected value of the random component and the systematic component.

I ηi =g(µi) whereµi=E(Yi)

and, g(.) is a continuous, monotone and differentiable function

I Identity link : g(µi) =µi

I Canonical link : g(µi) =θi

Regression Analysis III 8 / 14

(28)

Example

Example : Yi ∼Bernoulli(pi)

∴ ,µi =E(Yi) =pi

I logit link : g(µi) =log(1−µµi

i)

I probit link : g(µi) = Φ−1i)

I linear probability model : g(µi) =µi

(29)

Example

Example : Yi ∼Bernoulli(pi)

∴ ,µi =E(Yi) =pi

I logit link : g(µi) =log(1−µµi

i)

I probit link : g(µi) = Φ−1i)

I linear probability model : g(µi) =µi

Regression Analysis III 9 / 14

(30)

Example

Example : Yi ∼Bernoulli(pi)

∴ ,µi =E(Yi) =pi

I logit link : g(µi) =log(1−µµi

i)

I probit link : g(µi) = Φ−1i)

I linear probability model : g(µi) =µi

(31)

Example

Example : Yi ∼Bernoulli(pi)

∴ ,µi =E(Yi) =pi

I logit link : g(µi) =log(1−µµi

i)

I probit link : g(µi) = Φ−1i)

I linear probability model : g(µi) =µi

Regression Analysis III 9 / 14

(32)

Example

Example : Yi ∼Bernoulli(pi)

∴ ,µi =E(Yi) =pi

I logit link : g(µi) =log(1−µµi

i)

I probit link : g(µi) = Φ−1i)

I linear probability model : g(µi) =µi

(33)

Example

Example : Yi ∼Bernoulli(pi)

∴ ,µi =E(Yi) =pi

I logit link : g(µi) =log(1−µµi

i)

I probit link : g(µi) = Φ−1i)

I linear probability model : g(µi) =µi

Regression Analysis III 9 / 14

(34)

Likelihood estimation

log likelihood function based on a single observation :

I Li = yiθa(φi−b(θi)

i) +C(yi, φi)

I δLi

δθi = yi−ba(φ0i)

i)

I E(δLδθi

i) = 0⇒E(Yi) =µi =b0i)

I (δδθ2L2i i

) =−ba(φ00i)

i)

(35)

Likelihood estimation

log likelihood function based on a single observation :

I Li = yiθa(φi−b(θi)

i) +C(yi, φi)

I δLi

δθi = yi−ba(φ0i)

i)

I E(δLδθi

i) = 0⇒E(Yi) =µi =b0i)

I (δδθ2L2i i

) =−ba(φ00i)

i)

Regression Analysis III 10 / 14

(36)

Likelihood estimation

log likelihood function based on a single observation :

I Li = yiθa(φi−b(θi)

i) +C(yi, φi)

I δLi

δθi = yi−ba(φ0i)

i)

I E(δLδθi

i) = 0⇒E(Yi) =µi =b0i)

I (δδθ2L2i i

) =−ba(φ00i)

i)

(37)

Likelihood estimation

log likelihood function based on a single observation :

I Li = yiθa(φi−b(θi)

i) +C(yi, φi)

I δLi

δθi = yi−ba(φ0i)

i)

I E(δLδθi

i) = 0⇒E(Yi) =µi =b0i)

I (δδθ2L2i i

) =−ba(φ00i)

i)

Regression Analysis III 10 / 14

(38)

Likelihood estimation

log likelihood function based on a single observation :

I Li = yiθa(φi−b(θi)

i) +C(yi, φi)

I δLi

δθi = yi−ba(φ0i)

i)

I E(δLδθi

i) = 0⇒E(Yi) =µi =b0i)

I (δδθ2L2i i

) =−ba(φ00i)

i)

(39)

Variance function

E(δLδθi

i)2 =E[yi−ba(φ0i)

i) ]2

= V ar(Y[a(φ i)

i)]2

= ba(φ00i)

i)

⇒V ar(Yi) =b00i)a(φi)

V =b00i) : Variance function

V =V(µi) = dbd(θ0i)

i) = d(µd(θi)

i)

Regression Analysis III 11 / 14

(40)

Variance function

E(δLδθi

i)2 =E[yi−ba(φ0i)

i) ]2

= V ar(Y[a(φ i)

i)]2

= ba(φ00i)

i)

⇒V ar(Yi) =b00i)a(φi)

V =b00i) : Variance function

V =V(µ) = db0i) = d(µi)

(41)

Variance function

E(δLδθi

i)2 =E[yi−ba(φ0i)

i) ]2

= V ar(Y[a(φ i)

i)]2

= ba(φ00i)

i)

⇒V ar(Yi) =b00i)a(φi)

V =b00i) : Variance function

V =V(µi) = dbd(θ0i)

i) = d(µd(θi)

i)

Regression Analysis III 11 / 14

(42)

Variance function

E(δLδθi

i)2 =E[yi−ba(φ0i)

i) ]2

= V ar(Y[a(φ i)

i)]2

= ba(φ00i)

i)

⇒V ar(Yi) =b00i)a(φi)

V =b00i) : Variance function

V =V(µ) = db0i) = d(µi)

(43)

Salient features of GLM

I GLM does not transform Y, it only transforms the mean (E(Y)) and models it as a function of linear predictors.

I The objective is to investigate whether and how the mean varies as a function of the levels of our predictor or explanatory variables.

I Link function transforms the model to a linear model and retains the assumptions of normality with different mean for each observation Yi.

Regression Analysis III 12 / 14

(44)

Salient features of GLM

I GLM does not transform Y, it only transforms the mean (E(Y)) and models it as a function of linear predictors.

I The objective is to investigate whether and how the mean varies as a function of the levels of our predictor or explanatory variables.

I Link function transforms the model to a linear model and retains the assumptions of normality with different mean for each observation Yi.

(45)

Salient features of GLM

I GLM does not transform Y, it only transforms the mean (E(Y)) and models it as a function of linear predictors.

I The objective is to investigate whether and how the mean varies as a function of the levels of our predictor or explanatory variables.

I Link function transforms the model to a linear model and retains the assumptions of normality with different mean for each observation Yi.

Regression Analysis III 12 / 14

(46)

Salient features of GLM contd ...

I GLM relaxes normality.

I GLM allows for non-uniform variance.

I Variance of each observation Yi is a function of the meanµi.

I Distribution is completely specified in terms of its mean and variance.

(47)

Salient features of GLM contd ...

I GLM relaxes normality.

I GLM allows for non-uniform variance.

I Variance of each observation Yi is a function of the meanµi.

I Distribution is completely specified in terms of its mean and variance.

Regression Analysis III 13 / 14

(48)

Salient features of GLM contd ...

I GLM relaxes normality.

I GLM allows for non-uniform variance.

I Variance of each observation Yi is a function of the meanµi.

I Distribution is completely specified in terms of its mean and variance.

(49)

Salient features of GLM contd ...

I GLM relaxes normality.

I GLM allows for non-uniform variance.

I Variance of each observation Yi is a function of the meanµi.

I Distribution is completely specified in terms of its mean and variance.

Regression Analysis III 13 / 14

(50)

Summary

I Random component : the response variable Yi and it belongs to the one parameter exponential family.

I Systematic component : linear function of the explanatory variables (linear predictor).

I Link function : links the random component with the systematic component to make the relationship linear.

I Variance of each observation is a function of the mean of that observation.

(51)

Summary

I Random component : the response variable Yi and it belongs to the one parameter exponential family.

I Systematic component : linear function of the explanatory variables (linear predictor).

I Link function : links the random component with the systematic component to make the relationship linear.

I Variance of each observation is a function of the mean of that observation.

Regression Analysis III 14 / 14

(52)

Summary

I Random component : the response variable Yi and it belongs to the one parameter exponential family.

I Systematic component : linear function of the explanatory variables (linear predictor).

I Link function : links the random component with the systematic component to make the relationship linear.

I Variance of each observation is a function of the mean of that observation.

(53)

Summary

I Random component : the response variable Yi and it belongs to the one parameter exponential family.

I Systematic component : linear function of the explanatory variables (linear predictor).

I Link function : links the random component with the systematic component to make the relationship linear.

I Variance of each observation is a function of the mean of that observation.

Regression Analysis III 14 / 14

References

Related documents

I Categorical variable such as social class, educational level where the categories are ordered but the distance or spacing between the categories is not defined or unknown, is

I Our major objective while analyzing a contingency table is testing for independence of the response and predictor variable.. I For large samples we can use the Pearsonian

The International Model for Policy Analysis of Agricultural Commodities and Trade modeling system also analyzes the socioeconomic and climate pathways with different assumptions

One-way analysis variance for pharmacokinetic parameters and PROC GLM (procedure general linear model) followed by Dunnett's t and t test (parametric test) and

We used a cumulative link mixed-effects regression approach to model the transition of lantana from one qualitative density state to another as a function of biotic factors such

Based on the laboratory model test results for square foundation, an empirical nondimensional equation has been developed by regression analysis to determine the ultimate

Multivariate linear regression analysis method is a statistical technique for estimating the linear relationships among variables. It includes many techniques for

I It is more convenient to test the parameters of logit model than calculating many Odds Ratios from contingency tables and testing for them, when we have multiple predictor