对冲是什么意思| 公元500年是什么朝代| 糠是什么东西| bunny是什么意思| 肝郁脾虚吃什么药| 摩羯座属于什么象星座| 防微杜渐是什么意思| 什么妖魔鬼怪什么美女画皮| au990是什么金| 看心理医生挂什么科| top1是什么意思| 喝酒为什么会头疼| 手机cpu是什么| 柯萨奇病毒是什么病| 颈椎病挂号挂什么科| 1931年属相是什么| 水瓶座女生和什么星座男生最配| 什么伐桂| 孕激素高是什么原因| 肾虚是什么原因造成的| 独什么心什么| 维生素d3和d2有什么区别| 前胸后背出汗多是什么原因| 布偶猫长什么样| 两栖动物是什么意思| 脸为什么肿| 卡介苗为什么会留疤| 红眼病是什么原因引起的| 12月14号是什么星座| 脂肪由什么组成| 手术后吃什么| 蛋皮痒痒是什么病| 鹅和什么一起炖最好吃| 打飞机什么意思| 江与河有什么区别| 无字五行属什么| abo溶血症是什么意思| 天荒地老是什么生肖| 颈椎退行性病变是什么意思| 杨八妹属什么生肖| 4月17日是什么星座| 什么是有意义的东西| 什么叫荨麻疹| 宝宝便秘吃什么| 什么是湿热体质| 5月13日什么星座| 后妈是什么意思| 爱出汗的人是什么原因| 国色天香是什么生肖| 吃炒黄豆有什么好处和坏处| 什么地问填词语| 喝椰子汁有什么好处| 什么是提肛运动| cross是什么牌子| 黄金微针是什么| 老打瞌睡犯困是什么原因| 乳房胀痛是什么原因引起的| 口舌生疮吃什么药最好| 为什么咳嗽| 维生素h的作用及功能主治是什么| 吃什么鱼最健康| 腊梅什么时候开花| 什么有条| 打灰是什么意思| 日本打工需要什么条件| 哈伦裤配什么鞋子好看| 黄金的动物是什么生肖| 男龙和什么生肖最配| 子宫瘢痕是什么意思| 茶氨酸是什么| 缘是什么意思| 梅核气有什么症状| 凯莉包是什么牌子| 拉屎为什么是绿色的| 胚根发育成什么| 韭菜籽配什么壮阳最猛| 空调什么牌子的好| 血脂高有什么表现| 什么食物嘌呤高| 血糖高吃什么水果好能降糖| 唐筛检查什么| 二月初二是什么星座| 仪表堂堂是什么生肖| 叫人挪车打什么电话| 阿莫西林吃多了有什么副作用| 樱花的花语是什么| 眼睛痒用什么眼药水好| 乳腺癌ki67是什么意思| 什么食物降尿酸效果好| 四大才子是什么生肖| 装藏是什么意思| 肠胃炎吃什么水果比较好| 刚做了人流适合吃什么好| 大生化挂什么科| 十八岁成人礼送什么礼物| 什么叫体位性低血压| 男蛇配什么属相最好| 皮肤病挂什么科| 促进钙吸收吃什么| 口臭是什么原因| 医保和农村合作医疗有什么区别| bk病毒是什么| 禾加比读什么| 高胆固醇血症是什么意思| 黄瓜和什么一起炒好吃| 水疱疹什么药最快能治好| 咳嗽挂什么科| 画龙点睛指什么生肖| 月经前腰疼的厉害是什么原因| 颈椎反弓有什么症状| 移徙是什么意思| 霉菌阴道炎是什么引起的| 胃食管反流挂什么科| 红花有什么功效| 孕妇生气对胎儿有什么影响| 白手起家是什么生肖| 帮凶是什么意思| 望尘莫及什么意思| 拔完智齿吃什么消炎药| 阑尾炎痛起来什么感觉| 此物非彼物是什么意思| 高血压高血脂不能吃什么| 袁字五行属什么| 巨蟹男和什么座最配对| 胸闷气短吃什么药| 肺主皮毛是什么意思| 圣诞是什么意思| 手脚心发热是什么原因| 汽化是什么意思| 看见蛇有什么预兆| 男生适合什么职业| 男人血精是什么原因造成的| 纸鸢是什么意思| 吃什么补气血最快最好| 牛腩是什么部位的肉| 22年属什么生肖| 王母娘娘叫什么名字| 七月份生日是什么星座| 2月23是什么星座| 小苏打和食用碱有什么区别| 婴儿吃手是什么原因| 登高望远是什么生肖| epa是什么营养物质| 2月10号是什么星座| 心脏早搏吃什么药效果好| 只出不进什么意思| 女性潮红是什么意思| 甘油三酯偏高有什么危害| aurora是什么牌子| 贺涵为什么会爱上罗子君| 急性肠胃炎吃什么药| 手掌小鱼际发红是什么原因| 晴雨伞是什么意思| 青稞面是什么| 紫外线过敏用什么药| ivy什么意思| 锌补多了有什么症状| 朱元璋长什么样| 脂蛋白a高是什么原因| 什么止疼药见效最快| 蹲马步有什么好处| 为什么没有win9| hz是什么意思| 一直倒霉预示着什么| 希望孩子成为什么样的人| 心脏不舒服挂什么科室| 范思哲是什么品牌| 隆胸有什么危害和后遗症吗| 陈赫火锅店叫什么名字| 白马怕青牛是什么意思| 验尿能检查出什么| 黄五行属性是什么| 诗韵是什么意思| 前列腺炎吃什么消炎药好| 看望病人买什么东西好| 榴莲和什么相克| 儿童水杯什么材质好| 尿潴留是什么病| 女性胆囊炎有什么症状| 韦编三绝什么意思| 7月16是什么星座| 脑供血不足吃什么中成药好| 警惕是什么意思| hpd是什么意思| 发瘟是什么意思| 一国两制是什么时候提出的| 泌乳素高是什么原因| 晨勃是什么意思| 7月11是什么星座| 上午10点半是什么时辰| 天干是什么意思| 掂过碌蔗是什么意思| 贫血有什么危害| 天蝎座男和什么星座最配| 资金盘是什么意思| 家宴是什么意思| 东莞市委书记什么级别| 龙的幸运色是什么颜色| 什么是干咳| 头总是昏昏沉沉的是什么原因| 皮肤感染吃什么消炎药| 花甲和什么不能一起吃| 痛经吃什么食物| 流汗有什么好处| 什么水果含维生素b| 人嗜睡是什么原因| 口干口苦口臭是什么原因引起的| 蛋白石是什么石头| 牙齿遇冷热都痛是什么原因| 清影是什么意思| 世袭罔替什么意思| 氯偏高是什么原因| 一把把什么| 乳腺导管局限性扩张是什么意思| 姝姝是什么意思| 中性粒细胞偏低是什么原因| 什么的果子| 抗酸杆菌是什么意思| 包饺子剩下的面团能做什么| 集体户口和个人户口有什么区别| 怀孕初期头晕是什么原因| 姥姥的妈妈叫什么| 宝宝老是摇头是什么原因| 122是什么号码| 血钾是什么意思| 鸡肉和什么不能一起吃| 失责是什么意思| 葛根是什么| 痛经吃什么药最有效| 座山雕什么意思| 心肌缺血是什么原因| 护理考研考什么| 共振是什么意思| 高血压是什么引起的| 血稠吃什么药| 气管炎不能吃什么食物| 叶酸买什么牌子的好| 未见明显血流信号是什么意思| 小孩感冒发烧吃什么药| 老板喜欢什么样的员工| 枸杞泡水喝有什么作用和功效| 有什么花| 伐木是什么意思| 口若悬河是指什么生肖| 一什么杯子| 蚕豆病是什么病有什么症状| 什么案件才会出动便衣| 孕检无创是检查什么的| 勾践属什么生肖| 睾丸疼吃什么药| 3680是什么罩杯| 脚背上长痣代表什么| 什么是流程| 什么什么的阳光| 十二指肠球炎是什么病| 手指甲凹凸不平是什么原因| c919是什么意思| 生育酚乙酸酯是什么| 牛奶丝是什么面料| 牙齿根部发黑是什么原因| 急性肠胃炎可以吃什么| 优是什么意思| 静心什么意思| 手麻挂什么科最好| 百度

枣庄港航局召开航运污染防治暨重点项目调度会议

百度   与Uber一样要在自动驾驶时代发光发热的厂商还包括特斯拉、福特、通用和谷歌等大名鼎鼎的厂商。

In statistics, a generalized linear model (GLM) is a flexible generalization of ordinary linear regression. The GLM generalizes linear regression by allowing the linear model to be related to the response variable via a link function and by allowing the magnitude of the variance of each measurement to be a function of its predicted value.

Generalized linear models were formulated by John Nelder and Robert Wedderburn as a way of unifying various other statistical models, including linear regression, logistic regression and Poisson regression.[1] They proposed an iteratively reweighted least squares method for maximum likelihood estimation (MLE) of the model parameters. MLE remains popular and is the default method on many statistical computing packages. Other approaches, including Bayesian regression and least squares fitting to variance stabilized responses, have been developed.

Intuition

edit

Ordinary linear regression predicts the expected value of a given unknown quantity (the response variable, a random variable) as a linear combination of a set of observed values (predictors). This implies that a constant change in a predictor leads to a constant change in the response variable (i.e. a linear-response model). This is appropriate when the response variable can vary, to a good approximation, indefinitely in either direction, or more generally for any quantity that only varies by a relatively small amount compared to the variation in the predictive variables, e.g. human heights.

However, these assumptions are inappropriate for some types of response variables. For example, in cases where the response variable is expected to be always positive and varying over a wide range, constant input changes lead to geometrically (i.e. exponentially) varying, rather than constantly varying, output changes. As an example, suppose a linear prediction model learns from some data (perhaps primarily drawn from large beaches) that a 10 degree temperature decrease would lead to 1,000 fewer people visiting the beach. This model is unlikely to generalize well over differently-sized beaches. More specifically, the problem is that if the model is used to predict the new attendance with a temperature drop of 10 for a beach that regularly receives 50 beachgoers, it would predict an impossible attendance value of ?950. Logically, a more realistic model would instead predict a constant rate of increased beach attendance (e.g. an increase of 10 degrees leads to a doubling in beach attendance, and a drop of 10 degrees leads to a halving in attendance). Such a model is termed an exponential-response model (or log-linear model, since the logarithm of the response is predicted to vary linearly).

Similarly, a model that predicts a probability of making a yes/no choice (a Bernoulli variable) is even less suitable as a linear-response model, since probabilities are bounded on both ends (they must be between 0 and 1). Imagine, for example, a model that predicts the likelihood of a given person going to the beach as a function of temperature. A reasonable model might predict, for example, that a change in 10 degrees makes a person two times more or less likely to go to the beach. But what does "twice as likely" mean in terms of a probability? It cannot literally mean to double the probability value (e.g. 50% becomes 100%, 75% becomes 150%, etc.). Rather, it is the odds that are doubling: from 2:1 odds, to 4:1 odds, to 8:1 odds, etc. Such a model is a log-odds or logistic model.

Generalized linear models cover all these situations by allowing for response variables that have arbitrary distributions (rather than simply normal distributions), and for an arbitrary function of the response variable (the link function) to vary linearly with the predictors (rather than assuming that the response itself must vary linearly). For example, the case above of predicted number of beach attendees would typically be modeled with a Poisson distribution and a log link, while the case of predicted probability of beach attendance would typically be modelled with a Bernoulli distribution (or binomial distribution, depending on exactly how the problem is phrased) and a log-odds (or logit) link function.

Overview

edit

In a generalized linear model (GLM), each outcome Y of the dependent variables is assumed to be generated from a particular distribution in an exponential family, a large class of probability distributions that includes the normal, binomial, Poisson and gamma distributions, among others. The conditional mean μ of the distribution depends on the independent variables X through:

?

where E(Y?|?X) is the expected value of Y conditional on X; Xβ is the linear predictor, a linear combination of unknown parameters β; g is the link function.

In this framework, the variance is typically a function, V, of the mean:

?

It is convenient if V follows from an exponential family of distributions, but it may simply be that the variance is a function of the predicted value.

The unknown parameters, β, are typically estimated with maximum likelihood, maximum quasi-likelihood, or Bayesian techniques.

Model components

edit

The GLM consists of three elements:

1. A particular distribution for modeling ? from among those which are considered exponential families of probability distributions,
2. A linear predictor ?, and
3. A link function ? such that ?.

Probability distribution

edit

An overdispersed exponential family of distributions is a generalization of an exponential family and the exponential dispersion model of distributions and includes those families of probability distributions, parameterized by ? and ?, whose density functions f (or probability mass function, for the case of a discrete distribution) can be expressed in the form

?

The dispersion parameter, ?, typically is known and is usually related to the variance of the distribution. The functions ?, ?, ?, ?, and ? are known. Many common distributions are in this family, including the normal, exponential, gamma, Poisson, Bernoulli, and (for fixed number of trials) binomial, multinomial, and negative binomial.

For scalar ? and ? (denoted ? and ? in this case), this reduces to

?

? is related to the mean of the distribution. If ? is the identity function, then the distribution is said to be in canonical form (or natural form). Note that any distribution can be converted to canonical form by rewriting ? as ? and then applying the transformation ?. It is always possible to convert ? in terms of the new parametrization, even if ? is not a one-to-one function; see comments in the page on exponential families.

If, in addition, ? and ? are the identity, then ? is called the canonical parameter (or natural parameter) and is related to the mean through

?

For scalar ? and ?, this reduces to

?

Under this scenario, the variance of the distribution can be shown to be[2]

?

For scalar ? and ?, this reduces to

?

Linear predictor

edit

The linear predictor is the quantity which incorporates the information about the independent variables into the model. The symbol η (Greek "eta") denotes a linear predictor. It is related to the expected value of the data through the link function.

η is expressed as linear combinations (thus, "linear") of unknown parameters β. The coefficients of the linear combination are represented as the matrix of independent variables X. η can thus be expressed as

?
edit

The link function provides the relationship between the linear predictor and the mean of the distribution function. There are many commonly used link functions, and their choice is informed by several considerations. There is always a well-defined canonical link function which is derived from the exponential of the response's density function. However, in some cases it makes sense to try to match the domain of the link function to the range of the distribution function's mean, or use a non-canonical link function for algorithmic purposes, for example Bayesian probit regression.

When using a distribution function with a canonical parameter ? the canonical link function is the function that expresses ? in terms of ? i.e. ? For the most common distributions, the mean ? is one of the parameters in the standard form of the distribution's density function, and then ? is the function as defined above that maps the density function into its canonical form. When using the canonical link function, ? which allows ? to be a sufficient statistic for ?.

Following is a table of several exponential-family distributions in common use and the data they are typically used for, along with the canonical link functions and their inverses (sometimes referred to as the mean function, as done here).

Common distributions with typical uses and canonical link functions
Distribution Support of distribution Typical uses Link name Link function, ? Mean function
Normal real: ? Linear-response data Identity ? ?
Laplace
Exponential real: ? Exponential-response data, scale parameters Negative inverse ? ?
Gamma
Inverse
Gaussian
real: ? Inverse
squared
? ?
Poisson integer: ? count of occurrences in fixed amount of time/space Log ? ?
Bernoulli integer: ? outcome of single yes/no occurrence Logit ? ?
Binomial integer: ? count of # of "yes" occurrences out of N yes/no occurrences ?
Categorical integer: ? outcome of single K-way occurrence ?
K-vector of integer: ?, where exactly one element in the vector has the value 1
Multinomial K-vector of integer: ? count of occurrences of different types (1, ..., K) out of N total K-way occurrences

In the cases of the exponential and gamma distributions, the domain of the canonical link function is not the same as the permitted range of the mean. In particular, the linear predictor may be positive, which would give an impossible negative mean. When maximizing the likelihood, precautions must be taken to avoid this. An alternative is to use a noncanonical link function.

In the case of the Bernoulli, binomial, categorical and multinomial distributions, the support of the distributions is not the same type of data as the parameter being predicted. In all of these cases, the predicted parameter is one or more probabilities, i.e. real numbers in the range ?. The resulting model is known as logistic regression (or multinomial logistic regression in the case that K-way rather than binary values are being predicted).

For the Bernoulli and binomial distributions, the parameter is a single probability, indicating the likelihood of occurrence of a single event. The Bernoulli still satisfies the basic condition of the generalized linear model in that, even though a single outcome will always be either 0 or 1, the expected value will nonetheless be a real-valued probability, i.e. the probability of occurrence of a "yes" (or 1) outcome. Similarly, in a binomial distribution, the expected value is Np, i.e. the expected proportion of "yes" outcomes will be the probability to be predicted.

For categorical and multinomial distributions, the parameter to be predicted is a K-vector of probabilities, with the further restriction that all probabilities must add up to 1. Each probability indicates the likelihood of occurrence of one of the K possible values. For the multinomial distribution, and for the vector form of the categorical distribution, the expected values of the elements of the vector can be related to the predicted probabilities similarly to the binomial and Bernoulli distributions.

Fitting

edit

Maximum likelihood

edit

The maximum likelihood estimates can be found using an iteratively reweighted least squares algorithm or a Newton's method with updates of the form:

?

where ? is the observed information matrix (the negative of the Hessian matrix) and ? is the score function; or a Fisher's scoring method:

?

where ? is the Fisher information matrix. Note that if the canonical link function is used, then they are the same.[3]

Bayesian methods

edit

In general, the posterior distribution cannot be found in closed form and so must be approximated, usually using Laplace approximations or some type of Markov chain Monte Carlo method such as Gibbs sampling.

Examples

edit

General linear models

edit

A possible point of confusion has to do with the distinction between generalized linear models and general linear models, two broad statistical models. Co-originator John Nelder has expressed regret over this terminology.[4]

The general linear model may be viewed as a special case of the generalized linear model with identity link and responses normally distributed. As most exact results of interest are obtained only for the general linear model, the general linear model has undergone a somewhat longer historical development. Results for the generalized linear model with non-identity link are asymptotic (tending to work well with large samples).

Linear regression

edit

A simple, very important example of a generalized linear model (also an example of a general linear model) is linear regression. In linear regression, the use of the least-squares estimator is justified by the Gauss–Markov theorem, which does not assume that the distribution is normal.

From the perspective of generalized linear models, however, it is useful to suppose that the distribution function is the normal distribution with constant variance and the link function is the identity, which is the canonical link if the variance is known. Under these assumptions, the least-squares estimator is obtained as the maximum-likelihood parameter estimate.

For the normal distribution, the generalized linear model has a closed form expression for the maximum-likelihood estimates, which is convenient. Most other GLMs lack closed form estimates.

Binary data

edit

When the response data, Y, are binary (taking on only values 0 and 1), the distribution function is generally chosen to be the Bernoulli distribution and the interpretation of μi is then the probability, p, of Yi taking on the value one.

There are several popular link functions for binomial functions.

edit

The most typical link function is the canonical logit link:

?

GLMs with this setup are logistic regression models (or logit models).

edit

Alternatively, the inverse of any continuous cumulative distribution function (CDF) can be used for the link since the CDF's range is ?, the range of the binomial mean. The normal CDF ? is a popular choice and yields the probit model. Its link is

?

The reason for the use of the probit model is that a constant scaling of the input variable to a normal CDF (which can be absorbed through equivalent scaling of all of the parameters) yields a function that is practically identical to the logit function, but probit models are more tractable in some situations than logit models. (In a Bayesian setting in which normally distributed prior distributions are placed on the parameters, the relationship between the normal priors and the normal CDF link function means that a probit model can be computed using Gibbs sampling, while a logit model generally cannot.)

Complementary log-log (cloglog)

edit

The complementary log-log function may also be used:

?

This link function is asymmetric and will often produce different results from the logit and probit link functions.[5] The cloglog model corresponds to applications where we observe either zero events (e.g., defects) or one or more, where the number of events is assumed to follow the Poisson distribution.[6] The Poisson assumption means that

?

where μ is a positive number denoting the expected number of events. If p represents the proportion of observations with at least one event, its complement

?

and then

?

A linear model requires the response variable to take values over the entire real line. Since μ must be positive, we can enforce that by taking the logarithm, and letting log(μ) be a linear model. This produces the "cloglog" transformation

?
edit

The identity link g(p) = p is also sometimes used for binomial data to yield a linear probability model. However, the identity link can predict nonsense "probabilities" less than zero or greater than one. This can be avoided by using a transformation like cloglog, probit or logit (or any inverse cumulative distribution function). A primary merit of the identity link is that it can be estimated using linear math—and other standard link functions are approximately linear matching the identity link near p = 0.5.

Variance function

edit

The variance function for "quasibinomial" data is:

?

where the dispersion parameter τ is exactly 1 for the binomial distribution. Indeed, the standard binomial likelihood omits τ. When it is present, the model is called "quasibinomial", and the modified likelihood is called a quasi-likelihood, since it is not generally the likelihood corresponding to any real family of probability distributions. If τ exceeds 1, the model is said to exhibit overdispersion.

Multinomial regression

edit

The binomial case may be easily extended to allow for a multinomial distribution as the response (also, a Generalized Linear Model for counts, with a constrained total). There are two ways in which this is usually done:

Ordered response

edit

If the response variable is ordinal, then one may fit a model function of the form:

?

for m > 2. Different links g lead to ordinal regression models like proportional odds models or ordered probit models.

Unordered response

edit

If the response variable is a nominal measurement, or the data do not satisfy the assumptions of an ordered model, one may fit a model of the following form:

?

for m > 2. Different links g lead to multinomial logit or multinomial probit models. These are more general than the ordered response models, and more parameters are estimated.

Count data

edit

Another example of generalized linear models includes Poisson regression which models count data using the Poisson distribution. The link is typically the logarithm, the canonical link.

The variance function is proportional to the mean

?

where the dispersion parameter τ is typically fixed at exactly one. When it is not, the resulting quasi-likelihood model is often described as Poisson with overdispersion or quasi-Poisson.

Extensions

edit

Correlated or clustered data

edit

The standard GLM assumes that the observations are uncorrelated. Extensions have been developed to allow for correlation between observations, as occurs for example in longitudinal studies and clustered designs:

  • Generalized estimating equations (GEEs) allow for the correlation between observations without the use of an explicit probability model for the origin of the correlations, so there is no explicit likelihood. They are suitable when the random effects and their variances are not of inherent interest, as they allow for the correlation without explaining its origin. The focus is on estimating the average response over the population ("population-averaged" effects) rather than the regression parameters that would enable prediction of the effect of changing one or more components of X on a given individual. GEEs are usually used in conjunction with Huber–White standard errors.[7][8]
  • Generalized linear mixed models (GLMMs) are an extension to GLMs that includes random effects in the linear predictor, giving an explicit probability model that explains the origin of the correlations. The resulting "subject-specific" parameter estimates are suitable when the focus is on estimating the effect of changing one or more components of X on a given individual. GLMMs are also referred to as multilevel models and as mixed model. In general, fitting GLMMs is more computationally complex and intensive than fitting GEEs.

Generalized additive models

edit

Generalized additive models (GAMs) are another extension to GLMs in which the linear predictor η is not restricted to be linear in the covariates X but is the sum of smoothing functions applied to the xis:

?

The smoothing functions fi are estimated from the data. In general this requires a large number of data points and is computationally intensive.[9][10]

See also

edit

References

edit

Citations

edit
  1. ^ Nelder, John; Wedderburn, Robert (1972). "Generalized Linear Models". Journal of the Royal Statistical Society. Series A (General). 135 (3). Blackwell Publishing: 370–384. doi:10.2307/2344614. JSTOR?2344614. S2CID?14154576.
  2. ^ McCullagh & Nelder 1989, Chapter?2.
  3. ^ McCullagh & Nelder 1989, p.?43.
  4. ^ Senn, Stephen (2003). "A conversation with John Nelder". Statistical Science. 18 (1): 118–131. doi:10.1214/ss/1056397489. I suspect we should have found some more fancy name for it that would have stuck and not been confused with the general linear model, although general and generalized are not quite the same. I can see why it might have been better to have thought of something else.
  5. ^ "Complementary Log-log Model" (PDF).
  6. ^ "Which Link Function — Logit, Probit, or Cloglog?". Bayesium Analytics. 2025-08-14. Retrieved 2025-08-14.
  7. ^ Zeger, Scott L.; Liang, Kung-Yee; Albert, Paul S. (1988). "Models for Longitudinal Data: A Generalized Estimating Equation Approach". Biometrics. 44 (4). International Biometric Society: 1049–1060. doi:10.2307/2531734. JSTOR?2531734. PMID?3233245.
  8. ^ Hardin, James; Hilbe, Joseph (2003). Generalized Estimating Equations. London, England: Chapman and Hall/CRC. ISBN?1-58488-307-3.
  9. ^ Hastie & Tibshirani 1990.
  10. ^ Wood 2006.

Bibliography

edit

Further reading

edit
edit
为什么会长口腔溃疡 四十年婚姻是什么婚 脑梗是什么引起的 胃酸有什么办法缓解 孕妇吃什么好对胎儿好三个月前期
查甲状腺应该挂什么科 伤骨头了吃什么好得快 有一种水果叫什么竹 紫苏是什么 125是什么意思
儿童乐园有什么好玩的 男人是女人的什么 胰腺炎为什么不能同房 投射效应是什么意思 疝气是什么病怎样治疗
殇字是什么意思 胃复安又叫什么 7.17什么星座 luna什么意思 手脚发麻挂什么科
悬雍垂发炎吃什么药hcv7jop7ns4r.cn 喝什么养胃最好hcv7jop7ns2r.cn superman什么意思hcv9jop4ns2r.cn 胃肠感冒吃什么食物比较好hcv8jop4ns3r.cn 喉咙不舒服是什么原因hcv9jop5ns4r.cn
肌肉疼是什么原因hcv8jop4ns7r.cn 食古不化是什么意思zhongyiyatai.com 正因数是什么hcv8jop4ns4r.cn 肌酐低是什么意思啊hcv8jop9ns1r.cn 华为什么手机好dayuxmw.com
提肛运动有什么好处hcv7jop9ns3r.cn 檀木手串有什么好处hcv9jop4ns7r.cn 十八岁是什么年华inbungee.com 骨密度增高是什么意思hcv8jop8ns1r.cn 盐酸利多卡因注射作用是什么qingzhougame.com
黄丫头是什么鱼liaochangning.com 毁三观是什么意思啊cl108k.com 三道杠是什么牌子hcv7jop6ns8r.cn 42是什么生肖hcv8jop4ns4r.cn 他们吃什么hcv8jop8ns0r.cn
百度