• Nie Znaleziono Wyników

On Estimation of Bi‑liner Regression Parameters

N/A
N/A
Protected

Academic year: 2021

Share "On Estimation of Bi‑liner Regression Parameters"

Copied!
12
0
0

Pełen tekst

(1)

www.czasopisma.uni.lodz.pl/foe/ 6(332) 2017

[87]

Acta Universitatis Lodziensis

Folia Oeconomica

ISSN 0208-6018 e-ISSN 2353-7663

DOI: http://dx.doi.org/10.18778/0208‑6018.332.06

Grzegorz Antoni Sitek

University of Economics in Katowice, Faculty of Management, Department of Statistics, Econometrics and Mathematics, grzegors12@wp.pl 

Janusz Leszek Wywiał

University of Economics in Katowice, Faculty of Management, Department of Statistics, Econometrics and Mathematics, wywial@ue.katowice.pl

On Estimation of Bi‑liner Regression Parameters

Abstract: Two non‑parallel lines will be named as bi‑lines. The relationship between the definition

of the bi‑lines function and linear regression functions of the distribution mixture is considered. The bi‑lines function parameters are estimated using the least squares method for an implicit interde‑ pendence. In general, values of parameter estimators are evaluated by means of an approximation numerical method. In a particular case, the exact expressions for the parameter estimators were de‑ rived. In this particular case, the properties of the estimators are examined in details. The bi‑lines are also used to estimate the regression functions of the distribution mixture. The accuracy of the pa‑ rameter estimation is analyzed.

Keywords: bi‑lines function, linear regression, least squares method for an implicit interdependen‑

ce, mixture of probability distribution

(2)

88 Grzegorz Antoni Sitek, Janusz Leszek Wywiał

FOE 6(332) 2017 www.czasopisma.uni.lodz.pl/foe/

1. Introduction

The classic simple regression does not represent a two‑dimensional population if it is a bimodal population. Antoniewicz (1988; 2001) proposed to approximate probability distribution of a one‑dimensional random variable by means of two points. Generalizing this result he approximates a two‑dimensional distribution by means of two lines, a technique which he called “bi‑linear regression”. In sta-tistical literature, this term is rather used to define a specific linear model, see e.g. Gabriel (1998). In this paper, we will call Antoniewicz’s model simply “bi‑lines function”, because in general it leads to a two‑dimensional data spread approxima-tion by means of two lines. In the next part of this paper, the bi‑lines are compared with regressions of two‑dimensional probability mixture distributions, or more simply regressions of mixture distributions, see e.g. Quandt (1972). The mixture of Gaussian distributions has been extensively studied e.g. by McLachlan and Peel (2000) and Lindsay and Basak (1993).

The main purpose of the paper is finding differences between the bi‑lines function and regressions of mixture distributions. Moreover, we show that the es-timators of bi‑lines function parameters are biased eses-timators of appropriate pa-rameters of regressions of mixture distribution. In a particular case, the exact ex-pressions for the estimators of the bi‑lines function parameters, as well as for their variances will be derived.

Let a bivariate distribution of (X, Y) be a mixture of two bivariate distribu-tions of (X1, Y1) and (X2, Y2). Hence:

2 a specific linear model, see e.g. Gabriel (1998). In this paper, we will call Antoniewicz’s model simply “bi-lines function”, because in general it leads to a two-dimensional data spread approximation by means of two lines. In the next part of this paper, the bi-lines are compared with regressions of two-dimensional probability mixture distributions, or more simply regressions of mixture distributions, see e.g. Quandt (1972). The mixture of Gaussian distributions has been extensively studied e.g. by McLachlan and Peel (2000) and Lindsay and Basak (1993).

The main purpose of the paper is finding differences between the bi-lines function and regressions of mixture distributions. Moreover, we show that the estimators of bi-lines function parameters are biased estimators of appropriate parameters of regressions of mixture distribution. In a particular case, the exact expressions for the estimators of the bi-lines function parameters, as well as for their variances will be derived.

Let a bivariate distribution of (X, Y) be a mixture of two bivariate distributions of (X1, Y1)

and (X2, Y2). Hence: ) , ( ) 1 ( ) , ( ) , (x y pf1 x y p f2 x y f    , 0 < p < 1, (1)

where p is the mixing parameter. Let:

), ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) , ( ) (x f x y dy p f1 x y dy p f2 x y dy ph1 x p h2 x h

 

   (2) where: hi(x)

fi(x,y)dy, i = 1, 2, ). ( ) | ( ) ( ) | ( ) ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) | ( 1 1 2 2 2 1 2 1 f y x w x f y x w x x h p x ph y x f p y x pf x y f        (3) ) ( ) ( ) ( 1 1 x phh xx w  , . ) ( ) ( ) 1 ( ) ( 2 2 x hpxh x w  

It is obvious that if h1(x) = h2(x) then h(x) = h1(x) = h2(x) and w1(x) = p, w2(x) = 1 – p.

Moreover, in this case:

), | ( ) 1 ( ) | ( ) | (y x pf1 y x p f2 y x f    (4) ). | ( ) 1 ( ) | ( ) | (Y x pE Y1 x p E Y2 x E    (5)

Particularly, if h1(x) = h2(x) and E(Yi | x) = aix + bi, i = 1, 2, then expression (5) reduces to the

following: ). )( 1 ( ) ( ) | (Y x p a1x b1 p a2x b2 E      (6) (1) where p is the mixing parameter. Let:

2 a specific linear model, see e.g. Gabriel (1998). In this paper, we will call Antoniewicz’s model simply “bi-lines function”, because in general it leads to a two-dimensional data spread approximation by means of two lines. In the next part of this paper, the bi-lines are compared with regressions of two-dimensional probability mixture distributions, or more simply regressions of mixture distributions, see e.g. Quandt (1972). The mixture of Gaussian distributions has been extensively studied e.g. by McLachlan and Peel (2000) and Lindsay and Basak (1993).

The main purpose of the paper is finding differences between the bi-lines function and regressions of mixture distributions. Moreover, we show that the estimators of bi-lines function parameters are biased estimators of appropriate parameters of regressions of mixture distribution. In a particular case, the exact expressions for the estimators of the bi-lines function parameters, as well as for their variances will be derived.

Let a bivariate distribution of (X, Y) be a mixture of two bivariate distributions of (X1, Y1)

and (X2, Y2). Hence: ) , ( ) 1 ( ) , ( ) , (x y pf1 x y p f2 x y f    , 0 < p < 1, (1)

where p is the mixing parameter. Let:

), ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) , ( ) (x f x y dy p f1 x y dy p f2 x y dy ph1 x p h2 x h

 

   (2) where: hi(x)

fi(x,y)dy, i = 1, 2, ). ( ) | ( ) ( ) | ( ) ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) | ( 1 1 2 2 2 1 2 1 f y x w x f y x w x x h p x ph y x f p y x pf x y f        (3) ) ( ) ( ) ( 1 1 x phh xx w  , . ) ( ) ( ) 1 ( ) ( 2 2 x hpxh x w  

It is obvious that if h1(x) = h2(x) then h(x) = h1(x) = h2(x) and w1(x) = p, w2(x) = 1 – p.

Moreover, in this case:

), | ( ) 1 ( ) | ( ) | (y x pf1 y x p f2 y x f    (4) ). | ( ) 1 ( ) | ( ) | (Y x pE Y1 x p E Y2 x E    (5)

Particularly, if h1(x) = h2(x) and E(Yi | x) = aix + bi, i = 1, 2, then expression (5) reduces to the

following: ). )( 1 ( ) ( ) | (Y x p a1x b1 p a2x b2 E      (6) 2 a specific linear model, see e.g. Gabriel (1998). In this paper, we will call Antoniewicz’s model simply “bi-lines function”, because in general it leads to a two-dimensional data spread approximation by means of two lines. In the next part of this paper, the bi-lines are compared with regressions of two-dimensional probability mixture distributions, or more simply regressions of mixture distributions, see e.g. Quandt (1972). The mixture of Gaussian distributions has been extensively studied e.g. by McLachlan and Peel (2000) and Lindsay and Basak (1993).

The main purpose of the paper is finding differences between the bi-lines function and regressions of mixture distributions. Moreover, we show that the estimators of bi-lines function parameters are biased estimators of appropriate parameters of regressions of mixture distribution. In a particular case, the exact expressions for the estimators of the bi-lines function parameters, as well as for their variances will be derived.

Let a bivariate distribution of (X, Y) be a mixture of two bivariate distributions of (X1, Y1)

and (X2, Y2). Hence: ) , ( ) 1 ( ) , ( ) , (x y pf1 x y p f2 x y f    , 0 < p < 1, (1)

where p is the mixing parameter. Let:

), ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) , ( ) (x f x y dy p f1 x y dy p f2 x y dy ph1 x p h2 x h

 

   (2) where: hi(x)

fi(x,y)dy, i = 1, 2, ). ( ) | ( ) ( ) | ( ) ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) | ( 1 1 2 2 2 1 2 1 f y x w x f y x w x x h p x ph y x f p y x pf x y f        (3) ) ( ) ( ) ( 1 1 x phh xx w  , . ) ( ) ( ) 1 ( ) ( 2 2 x hpxh x w  

It is obvious that if h1(x) = h2(x) then h(x) = h1(x) = h2(x) and w1(x) = p, w2(x) = 1 – p.

Moreover, in this case:

), | ( ) 1 ( ) | ( ) | (y x pf1 y x p f2 y x f    (4) ). | ( ) 1 ( ) | ( ) | (Y x pE Y1 x p E Y2 x E    (5)

Particularly, if h1(x) = h2(x) and E(Yi | x) = aix + bi, i = 1, 2, then expression (5) reduces to the

following: ). )( 1 ( ) ( ) | (Y x p a1x b1 p a2x b2 E      (6) (2) where: 2 a specific linear model, see e.g. Gabriel (1998). In this paper, we will call Antoniewicz’s model simply “bi-lines function”, because in general it leads to a two-dimensional data spread approximation by means of two lines. In the next part of this paper, the bi-lines are compared with regressions of two-dimensional probability mixture distributions, or more simply regressions of mixture distributions, see e.g. Quandt (1972). The mixture of Gaussian distributions has been extensively studied e.g. by McLachlan and Peel (2000) and Lindsay and Basak (1993).

The main purpose of the paper is finding differences between the bi-lines function and regressions of mixture distributions. Moreover, we show that the estimators of bi-lines function parameters are biased estimators of appropriate parameters of regressions of mixture distribution. In a particular case, the exact expressions for the estimators of the bi-lines function parameters, as well as for their variances will be derived.

Let a bivariate distribution of (X, Y) be a mixture of two bivariate distributions of (X1, Y1)

and (X2, Y2). Hence: ) , ( ) 1 ( ) , ( ) , (x y pf1 x y p f2 x y f    , 0 < p < 1, (1)

where p is the mixing parameter. Let:

), ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) , ( ) (x f x y dy p f1 x y dy p f2 x y dy ph1 x p h2 x h

 

   (2) where: hi(x)

fi(x,y)dy, i = 1, 2, ). ( ) | ( ) ( ) | ( ) ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) | ( 1 1 2 2 2 1 2 1 f y x w x f y x w x x h p x ph y x f p y x pf x y f        (3) ) ( ) ( ) ( 1 1 x phh xx w  , . ) ( ) ( ) 1 ( ) ( 2 2 x hpxh x w  

It is obvious that if h1(x) = h2(x) then h(x) = h1(x) = h2(x) and w1(x) = p, w2(x) = 1 – p.

Moreover, in this case:

), | ( ) 1 ( ) | ( ) | (y x pf1 y x p f2 y x f    (4) ). | ( ) 1 ( ) | ( ) | (Y x pE Y1 x p E Y2 x E    (5)

Particularly, if h1(x) = h2(x) and E(Yi | x) = aix + bi, i = 1, 2, then expression (5) reduces to the

following: ). )( 1 ( ) ( ) | (Y x p a1x b1 p a2x b2 E      (6) , i = 1, 2, 2 a specific linear model, see e.g. Gabriel (1998). In this paper, we will call Antoniewicz’s model simply “bi-lines function”, because in general it leads to a two-dimensional data spread approximation by means of two lines. In the next part of this paper, the bi-lines are compared with regressions of two-dimensional probability mixture distributions, or more simply regressions of mixture distributions, see e.g. Quandt (1972). The mixture of Gaussian distributions has been extensively studied e.g. by McLachlan and Peel (2000) and Lindsay and Basak (1993).

The main purpose of the paper is finding differences between the bi-lines function and regressions of mixture distributions. Moreover, we show that the estimators of bi-lines function parameters are biased estimators of appropriate parameters of regressions of mixture distribution. In a particular case, the exact expressions for the estimators of the bi-lines function parameters, as well as for their variances will be derived.

Let a bivariate distribution of (X, Y) be a mixture of two bivariate distributions of (X1, Y1)

and (X2, Y2). Hence: ) , ( ) 1 ( ) , ( ) , (x y pf1 x y p f2 x y f    , 0 < p < 1, (1)

where p is the mixing parameter. Let:

), ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) , ( ) (x f x y dy p f1 x y dy p f2 x y dy ph1 x p h2 x h

 

   (2) where: hi(x)

fi(x,y)dy, i = 1, 2, ). ( ) | ( ) ( ) | ( ) ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) | ( 1 1 2 2 2 1 2 1 f y x w x f y x w x x h p x ph y x f p y x pf x y f        (3) ) ( ) ( ) ( 1 1 x phh xx w  , . ) ( ) ( ) 1 ( ) ( 2 2 x hpxh x w  

It is obvious that if h1(x) = h2(x) then h(x) = h1(x) = h2(x) and w1(x) = p, w2(x) = 1 – p.

Moreover, in this case:

), | ( ) 1 ( ) | ( ) | (y x pf1 y x p f2 y x f    (4) ). | ( ) 1 ( ) | ( ) | (Y x pE Y1 x p E Y2 x E    (5)

Particularly, if h1(x) = h2(x) and E(Yi | x) = aix + bi, i = 1, 2, then expression (5) reduces to the

following: ). )( 1 ( ) ( ) | (Y x p a1x b1 p a2x b2 E      (6) (3) 2 a specific linear model, see e.g. Gabriel (1998). In this paper, we will call Antoniewicz’s model simply “bi-lines function”, because in general it leads to a two-dimensional data spread approximation by means of two lines. In the next part of this paper, the bi-lines are compared with regressions of two-dimensional probability mixture distributions, or more simply regressions of mixture distributions, see e.g. Quandt (1972). The mixture of Gaussian distributions has been extensively studied e.g. by McLachlan and Peel (2000) and Lindsay and Basak (1993).

The main purpose of the paper is finding differences between the bi-lines function and regressions of mixture distributions. Moreover, we show that the estimators of bi-lines function parameters are biased estimators of appropriate parameters of regressions of mixture distribution. In a particular case, the exact expressions for the estimators of the bi-lines function parameters, as well as for their variances will be derived.

Let a bivariate distribution of (X, Y) be a mixture of two bivariate distributions of (X1, Y1)

and (X2, Y2). Hence: ) , ( ) 1 ( ) , ( ) , (x y pf1 x y p f2 x y f    , 0 < p < 1, (1)

where p is the mixing parameter. Let:

), ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) , ( ) (x f x y dy p f1 x y dy p f2 x y dy ph1 x p h2 x h

 

   (2) where: hi(x)

fi(x,y)dy, i = 1, 2, ). ( ) | ( ) ( ) | ( ) ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) | ( 1 1 2 2 2 1 2 1 f y x w x f y x w x x h p x ph y x f p y x pf x y f        (3) ) ( ) ( ) ( 1 1 x h x ph x w  , . ) ( ) ( ) 1 ( ) ( 2 2 x hpxh x w  

It is obvious that if h1(x) = h2(x) then h(x) = h1(x) = h2(x) and w1(x) = p, w2(x) = 1 – p.

Moreover, in this case:

), | ( ) 1 ( ) | ( ) | (y x pf1 y x p f2 y x f    (4) ). | ( ) 1 ( ) | ( ) | (Y x pE Y1 x p E Y2 x E    (5)

Particularly, if h1(x) = h2(x) and E(Yi | x) = aix + bi, i = 1, 2, then expression (5) reduces to the

following: ). )( 1 ( ) ( ) | (Y x p a1x b1 p a2x b2 E      (6)

(3)

On Estimation of Bi‑liner Regression Parameters 89

www.czasopisma.uni.lodz.pl/foe/ FOE 6(332) 2017 It is obvious that if h1(x) = h2(x) then h(x) = h1(x) = h2(x) and w1(x) = p,

w2(x) = 1 – p. Moreover, in this case:

2 a specific linear model, see e.g. Gabriel (1998). In this paper, we will call Antoniewicz’s model simply “bi-lines function”, because in general it leads to a two-dimensional data spread approximation by means of two lines. In the next part of this paper, the bi-lines are compared with regressions of two-dimensional probability mixture distributions, or more simply regressions of mixture distributions, see e.g. Quandt (1972). The mixture of Gaussian distributions has been extensively studied e.g. by McLachlan and Peel (2000) and Lindsay and Basak (1993).

The main purpose of the paper is finding differences between the bi-lines function and regressions of mixture distributions. Moreover, we show that the estimators of bi-lines function parameters are biased estimators of appropriate parameters of regressions of mixture distribution. In a particular case, the exact expressions for the estimators of the bi-lines function parameters, as well as for their variances will be derived.

Let a bivariate distribution of (X, Y) be a mixture of two bivariate distributions of (X1, Y1)

and (X2, Y2). Hence: ) , ( ) 1 ( ) , ( ) , (x y pf1 x y p f2 x y f    , 0 < p < 1, (1)

where p is the mixing parameter. Let:

), ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) , ( ) (x f x y dy p f1 x y dy p f2 x y dy ph1 x p h2 x h

 

   (2) where: hi(x)

fi(x,y)dy, i = 1, 2, ). ( ) | ( ) ( ) | ( ) ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) | ( 1 1 2 2 2 1 2 1 f y x w x f y x w x x h p x ph y x f p y x pf x y f        (3) ) ( ) ( ) ( 1 1 x h x ph x w  , . ) ( ) ( ) 1 ( ) ( 2 2 x hpxh x w  

It is obvious that if h1(x) = h2(x) then h(x) = h1(x) = h2(x) and w1(x) = p, w2(x) = 1 – p.

Moreover, in this case:

), | ( ) 1 ( ) | ( ) | (y x pf1 y x p f2 y x f    (4) ). | ( ) 1 ( ) | ( ) | (Y x pE Y1 x p E Y2 x E    (5)

Particularly, if h1(x) = h2(x) and E(Yi | x) = aix + bi, i = 1, 2, then expression (5) reduces to the

following: ). )( 1 ( ) ( ) | (Y x p a1x b1 p a2x b2 E      (6) (4) 2 a specific linear model, see e.g. Gabriel (1998). In this paper, we will call Antoniewicz’s model simply “bi-lines function”, because in general it leads to a two-dimensional data spread approximation by means of two lines. In the next part of this paper, the bi-lines are compared with regressions of two-dimensional probability mixture distributions, or more simply regressions of mixture distributions, see e.g. Quandt (1972). The mixture of Gaussian distributions has been extensively studied e.g. by McLachlan and Peel (2000) and Lindsay and Basak (1993).

The main purpose of the paper is finding differences between the bi-lines function and regressions of mixture distributions. Moreover, we show that the estimators of bi-lines function parameters are biased estimators of appropriate parameters of regressions of mixture distribution. In a particular case, the exact expressions for the estimators of the bi-lines function parameters, as well as for their variances will be derived.

Let a bivariate distribution of (X, Y) be a mixture of two bivariate distributions of (X1, Y1)

and (X2, Y2). Hence: ) , ( ) 1 ( ) , ( ) , (x y pf1 x y p f2 x y f    , 0 < p < 1, (1)

where p is the mixing parameter. Let:

), ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) , ( ) (x f x y dy p f1 x y dy p f2 x y dy ph1 x p h2 x h

 

   (2) where: hi(x)

fi(x,y)dy, i = 1, 2, ). ( ) | ( ) ( ) | ( ) ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) | ( 1 1 2 2 2 1 2 1 f y x w x f y x w x x h p x ph y x f p y x pf x y f        (3) ) ( ) ( ) ( 1 1 x phh xx w  , . ) ( ) ( ) 1 ( ) ( 2 2 x hpxh x w  

It is obvious that if h1(x) = h2(x) then h(x) = h1(x) = h2(x) and w1(x) = p, w2(x) = 1 – p.

Moreover, in this case:

), | ( ) 1 ( ) | ( ) | (y x pf1 y x p f2 y x f    (4) ). | ( ) 1 ( ) | ( ) | (Y x pEY1 x p E Y2 x E    (5)

Particularly, if h1(x) = h2(x) and E(Yi | x) = aix + bi, i = 1, 2, then expression (5) reduces to the

following: ). )( 1 ( ) ( ) | (Y x p a1x b1 p a2x b2 E      (6) (5)

Particularly, if h1(x) = h2(x) and E(Yi | x) = aix + bi, i = 1, 2, then expression (5) reduces to the following:

2 a specific linear model, see e.g. Gabriel (1998). In this paper, we will call Antoniewicz’s model simply “bi-lines function”, because in general it leads to a two-dimensional data spread approximation by means of two lines. In the next part of this paper, the bi-lines are compared with regressions of two-dimensional probability mixture distributions, or more simply regressions of mixture distributions, see e.g. Quandt (1972). The mixture of Gaussian distributions has been extensively studied e.g. by McLachlan and Peel (2000) and Lindsay and Basak (1993).

The main purpose of the paper is finding differences between the bi-lines function and regressions of mixture distributions. Moreover, we show that the estimators of bi-lines function parameters are biased estimators of appropriate parameters of regressions of mixture distribution. In a particular case, the exact expressions for the estimators of the bi-lines function parameters, as well as for their variances will be derived.

Let a bivariate distribution of (X, Y) be a mixture of two bivariate distributions of (X1, Y1)

and (X2, Y2). Hence: ) , ( ) 1 ( ) , ( ) , (x y pf1 x y p f2 x y f    , 0 < p < 1, (1)

where p is the mixing parameter. Let:

), ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) , ( ) (x f x y dy p f1 x y dy p f2 x y dy ph1 x p h2 x h

 

   (2) where: hi(x)

fi(x,y)dy, i = 1, 2, ). ( ) | ( ) ( ) | ( ) ( ) 1 ( ) ( ) , ( ) 1 ( ) , ( ) | ( 1 1 2 2 2 1 2 1 f y x w x f y x w x x h p x ph y x f p y x pf x y f        (3) ) ( ) ( ) ( 1 1 x h x ph x w  , . ) ( ) ( ) 1 ( ) ( 2 2 x h x h p x w  

It is obvious that if h1(x) = h2(x) then h(x) = h1(x) = h2(x) and w1(x) = p, w2(x) = 1 – p.

Moreover, in this case:

), | ( ) 1 ( ) | ( ) | (y x pf1 y x p f2 y x f    (4) ). | ( ) 1 ( ) | ( ) | (Y x pE Y1 x p E Y2 x E    (5)

Particularly, if h1(x) = h2(x) and E(Yi | x) = aix + bi, i = 1, 2, then expression (5) reduces to the

following: ). )( 1 ( ) ( ) | (Y x p a1x b1 p a2x b2 E      (6) (6)

The above function we will treat as a linear regression function of the mix-ture distribution while E(Yi | x) = aix + bi, i = 1, 2, are linear regression functions of the mixture distribution.

2. The least squares method for an implicit

interdependence

Let (X, Y) be a two‑dimensional random variable. Antoniewicz (1988) proposed an original method of approximation distribution by means of two lines, neither of which is parallel to the axis of the system. Parameters a, b, c and d of the lines minimize the following function:

The above function we will treat as a linear regression function of the mixture distribution while E(Yi | x) = aix + bi, i = 1, 2, are linear regression functions of the mixture distribution.

2. The least squares method for an implicit interdependence

Let (X, Y) be a two-dimensional random variable. Antoniewicz (1988) proposed an original method of approximation distribution by means of two lines, neither of which is parallel to the axis of the system. Parameters a, b, c and d of the lines minimize the following function:

 

. ) , , , (a EYaX c 2YbXd 2     (7)

Based on the available data, the parameters a, b, c and d will be estimated. This is equivalent to finding the straight lines that give the best fit (representation) of the points in the scatter plot of the response versus the predictor variable. We estimate the parameters using the popular least squares method, which gives the lines that minimize the sum of squares of the vertical distances from each point to the lines. The vertical distances represent the errors in the response variable. The sum of squares of these distances can then be written as follows: ( , , , ) [( )( )]. 1 2

        n i i i i i d x b y c x a y d c b a S (8)

The values of estimators aˆ,bcdˆ which minimize S(a, b, c, d) are derived by Antoniewicz (1988). In general, values of aˆ,bcdˆ can be obtained only numerically. This problem is considered in details by Sitek (2016).

If c = d = 0, expression (8) reduces to the following: . ] ) )( [( ) , ( 1 2

    n i i i i i bx y ax y b a S (9)

Values of a , that minimize S(a, b) are given by the solution (roots) of the following ^, b^ nonlinear system of two equations (10).

               0 2 2 0 2 2 40 2 31 22 31 2 22 13 40 2 31 22 31 2 22 13 bm a abm bm m a am m m ab abm am m b bm m (10) where

  n i v i u i uv x y m 1 , u = 0, 1, 2, …, v = 0, 1, 2, …

Under the assumption that a ≠ b, after appropriate transformations we have:

(7) Based on the available data, the parameters a, b, c and d will be estimated. This is equivalent to finding the straight lines that give the best fit (representation) of the points in the scatter plot of the response versus the predictor variable. We es-timate the parameters using the popular least squares method, which gives the lines that minimize the sum of squares of the vertical distances from each point to the lines. The vertical distances represent the errors in the response variable. The sum of squares of these distances can then be written as follows:

3 The above function we will treat as a linear regression function of the mixture distribution while E(Yi | x) = aix + bi, i = 1, 2, are linear regression functions of the mixture distribution.

2. The least squares method for an implicit interdependence

Let (X, Y) be a two-dimensional random variable. Antoniewicz (1988) proposed an original method of approximation distribution by means of two lines, neither of which is parallel to the axis of the system. Parameters a, b, c and d of the lines minimize the following function:

 

. ) , , , (a EYaX c2YbX d 2     (7)

Based on the available data, the parameters a, b, c and d will be estimated. This is equivalent to finding the straight lines that give the best fit (representation) of the points in the scatter plot of the response versus the predictor variable. We estimate the parameters using the popular least squares method, which gives the lines that minimize the sum of squares of the vertical distances from each point to the lines. The vertical distances represent the errors in the response variable. The sum of squares of these distances can then be written as follows: ( , , , ) [( )( )] . 1 2

        n i i i i i d x b y c x a y d c b a S (8)

The values of estimators aˆ,bcdˆ which minimize S(a, b, c, d) are derived by Antoniewicz (1988). In general, values of aˆ,bcdˆ can be obtained only numerically. This problem is considered in details by Sitek (2016).

If c = d = 0, expression (8) reduces to the following: . ] ) )( [( ) , ( 1 2

    n i i i i i bx y ax y b a S (9)

Values of a , that minimize S(a, b) are given by the solution (roots) of the following ^, b^ nonlinear system of two equations (10).

               0 2 2 0 2 2 40 2 31 22 31 2 22 13 40 2 31 22 31 2 22 13 bm a abm bm m a am m m ab abm am m b bm m (10) where

  n i v i u i uv x y m 1 , u = 0, 1, 2, …, v = 0, 1, 2, …

Under the assumption that a ≠ b, after appropriate transformations we have:

(4)

90 Grzegorz Antoni Sitek, Janusz Leszek Wywiał

FOE 6(332) 2017 www.czasopisma.uni.lodz.pl/foe/

The values of estimators abcdˆ which minimize S(a, b, c, d) are derived by Antoniewicz (1988). In general, values of abcdˆ can be obtained only nu-merically. This problem is considered in details by Sitek (2016).

If c = d = 0, expression (8) reduces to the following:

3 The above function we will treat as a linear regression function of the mixture distribution while E(Yi | x) = aix + bi, i = 1, 2, are linear regression functions of the mixture distribution.

2. The least squares method for an implicit interdependence

Let (X, Y) be a two-dimensional random variable. Antoniewicz (1988) proposed an original method of approximation distribution by means of two lines, neither of which is parallel to the axis of the system. Parameters a, b, c and d of the lines minimize the following function:

 

. ) , , , (a EYaXc2 YbXd 2     (7)

Based on the available data, the parameters a, b, c and d will be estimated. This is equivalent to finding the straight lines that give the best fit (representation) of the points in the scatter plot of the response versus the predictor variable. We estimate the parameters using the popular least squares method, which gives the lines that minimize the sum of squares of the vertical distances from each point to the lines. The vertical distances represent the errors in the response variable. The sum of squares of these distances can then be written as follows: ( , , , ) [( )( )] . 1 2

        n i i i i i d x b y c x a y d c b a S (8)

The values of estimators aˆ,bcdˆ which minimize S(a, b, c, d) are derived by Antoniewicz (1988). In general, values of aˆ,bcdˆ can be obtained only numerically. This problem is considered in details by Sitek (2016).

If c = d = 0, expression (8) reduces to the following: . ] ) )( [( ) , ( 1 2

    n i i i i i bx y ax y b a S (9)

Values of a , that minimize S(a, b) are given by the solution (roots) of the following ^, b^ nonlinear system of two equations (10).

               0 2 2 0 2 2 40 2 31 22 31 2 22 13 40 2 31 22 31 2 22 13 bm a abm bm m a am m m ab abm am m b bm m (10) where

  n i v i u i uv x y m 1 , u = 0, 1, 2, …, v = 0, 1, 2, …

Under the assumption that a ≠ b, after appropriate transformations we have:

(9)

Values of a^, b^, that minimize S(a, b) are given by the solution (roots) of the following nonlinear system of two equations (10).

3 The above function we will treat as a linear regression function of the mixture distribution while E(Yi | x) = aix + bi, i = 1, 2, are linear regression functions of the mixture distribution.

2. The least squares method for an implicit interdependence

Let (X, Y) be a two-dimensional random variable. Antoniewicz (1988) proposed an original method of approximation distribution by means of two lines, neither of which is parallel to the axis of the system. Parameters a, b, c and d of the lines minimize the following function:

 

. ) , , , (a EYaXc2 YbXd 2     (7)

Based on the available data, the parameters a, b, c and d will be estimated. This is equivalent to finding the straight lines that give the best fit (representation) of the points in the scatter plot of the response versus the predictor variable. We estimate the parameters using the popular least squares method, which gives the lines that minimize the sum of squares of the vertical distances from each point to the lines. The vertical distances represent the errors in the response variable. The sum of squares of these distances can then be written as follows: ( , , , ) [( )( )] . 1 2

        n i i i i i d x b y c x a y d c b a S (8)

The values of estimators aˆ,bcdˆ which minimize S(a, b, c, d) are derived by Antoniewicz (1988). In general, values of aˆ,bcdˆ can be obtained only numerically. This problem is considered in details by Sitek (2016).

If c = d = 0, expression (8) reduces to the following: . ] ) )( [( ) , ( 1 2

    n i i i i i bx y ax y b a S (9)

Values of a , that minimize S(a, b) are given by the solution (roots) of the following ^, b^ nonlinear system of two equations (10).

               0 2 2 0 2 2 40 2 31 22 31 2 22 13 40 2 31 22 31 2 22 13 bm a abm bm m a am m m ab abm am m b bm m (10) where

  n i v i u i uv x y m 1 , u = 0, 1, 2, …, v = 0, 1, 2, …

Under the assumption that a ≠ b, after appropriate transformations we have:

(10)

where

3 The above function we will treat as a linear regression function of the mixture distribution while E(Yi | x) = aix + bi, i = 1, 2, are linear regression functions of the mixture distribution.

2. The least squares method for an implicit interdependence

Let (X, Y) be a two-dimensional random variable. Antoniewicz (1988) proposed an original method of approximation distribution by means of two lines, neither of which is parallel to the axis of the system. Parameters a, b, c and d of the lines minimize the following function:

 

. ) , , , (a EYaXc2 YbXd 2     (7)

Based on the available data, the parameters a, b, c and d will be estimated. This is equivalent to finding the straight lines that give the best fit (representation) of the points in the scatter plot of the response versus the predictor variable. We estimate the parameters using the popular least squares method, which gives the lines that minimize the sum of squares of the vertical distances from each point to the lines. The vertical distances represent the errors in the response variable. The sum of squares of these distances can then be written as follows: ( , , , ) [( )( )] . 1 2

        n i i i i i d x b y c x a y d c b a S (8)

The values of estimators aˆ,bcdˆ which minimize S(a, b, c, d) are derived by Antoniewicz (1988). In general, values of aˆ,bcdˆ can be obtained only numerically. This problem is considered in details by Sitek (2016).

If c = d = 0, expression (8) reduces to the following: . ] ) )( [( ) , ( 1 2

    n i i i i i bx y ax y b a S (9)

Values of a , that minimize S(a, b) are given by the solution (roots) of the following ^, b^ nonlinear system of two equations (10).

               0 2 2 0 2 2 40 2 31 22 31 2 22 13 40 2 31 22 31 2 22 13 bm a abm bm m a am m m ab abm am m b bm m (10) where

  n i v i u i uv x y m 1 , u = 0, 1, 2, …, v = 0, 1, 2, …

Under the assumption that a ≠ b, after appropriate transformations we have:Under the assumption that a ≠ b, after appropriate transformations we have: ) /( 0 ) ( ) ( ) (ba m22 b2a2 m31abba m40  ba  40 31 31 22 bm m bm m a    . (11) (12)

After putting the right site of equation (11) into the first equation of system (10) we obtain the following quadratic equation,

. 0 ) ( ) ( 2 22 13 31 40 13 22 31 2 31 22 40 2 m m m b m m m m m m m b (12) Next, we have:

...

( ) 4( )( 2). 22 13 31 2 31 22 40 2 40 13 22 31m m m m m m m m m m muv        

If Δ > 0 then there are the following two distinct roots:

, ) ( 2 ) )( ( 4 ) ( ˆ 40 22 2 31 13 31 2 22 22 40 2 31 2 40 31 22 13 40 13 31 22 1 m m m m m m mm mm m m m m m m m b         (13) . ) ( 2 ) )( ( 4 ) ( ˆ 40 22 2 31 13 31 2 22 22 40 2 31 2 40 31 22 13 40 13 31 22 2 m m m m m m m m m m m m m m m m m b         (14)

Substituting the results into the equation (12), we get: . ˆ ˆ ˆ 40 1 31 31 1 22 1 m b m m b m a   

Using the well-known Vieta’s formulas, it can be shown that a  : ˆ b1 ˆ2 ), ˆ ( / ˆ ˆ ˆ 40 1 31 2 40 1 31 31 1 22 b m bm m b m m b m   ), ˆ ( ˆ ˆ 40 1 31 2 31 1 22 bm b m bm m     , ˆ ˆ ˆ ˆ 40 1 2 31 2 31 1 22 bm b m bbm m     . ) ˆ ˆ ( ˆ ˆ 31 2 1 40 1 2 22 bbm b b m m    

Let us introduce notation:

 

2

31 22 40m m m m A Auv   , BB

 

muvm31m22m13m40,

 

2 22 13 31m m m m C Cuv   .

Hence, from Viete’s formula we have:

) /( 0 ) ( ) ( ) ( 2 2 31 40 22 b a m ab b a m b a m a b        40 31 31 22 bm m bm m a    . (11) (12)

After putting the right site of equation (11) into the first equation of system (10) we obtain the following quadratic equation,

. 0 ) ( ) ( 2 22 13 31 40 13 22 31 2 31 22 40 2 m m m bm m m m m m m b (12) Next, we have:

...

( ) 4( )( 2). 22 13 31 2 31 22 40 2 40 13 22 31m m m m m m m m m m muv        

If Δ > 0 then there are the following two distinct roots:

, ) ( 2 ) )( ( 4 ) ( ˆ 40 22 2 31 13 31 2 22 22 40 2 31 2 40 31 22 13 40 13 31 22 1 m m m m m m m m m m m m m m m m m b         (13) . ) ( 2 ) )( ( 4 ) ( ˆ 40 22 2 31 13 31 2 22 22 40 2 31 2 40 31 22 13 40 13 31 22 2 m m m m m m m m m m m m m m m m m b         (14)

Substituting the results into the equation (12), we get: . ˆ ˆ ˆ 40 1 31 31 1 22 1 m b m m b m a   

Using the well-known Vieta’s formulas, it can be shown that a  : ˆ b1 ˆ2 ), ˆ ( / ˆ ˆ ˆ 40 1 31 2 40 1 31 31 1 22 b m bm m b m m b m   ), ˆ ( ˆ ˆ1 31 2 31 1 40 22 bm b m bm m     , ˆ ˆ ˆ ˆ1 31 2 31 2 1 40 22 bm b m bbm m     . ) ˆ ˆ ( ˆ ˆ 31 2 1 40 1 2 22 bbm b b m m    

Let us introduce notation:

 

2

31 22 40m m m m A Auv   , BB

 

muvm31m22m13m40,

 

2 22 13 31m m m m C Cuv   .

Hence, from Viete’s formula we have:

(11) After putting the right site of equation (11) into the first equation of system (10) we obtain the following quadratic equation,

4 ) /( 0 ) ( ) ( ) ( 2 2 31 40 22 b a m abb a m b a m a b        40 31 31 22 bm m bm m a    . (11) (12)

After putting the right site of equation (11) into the first equation of system (10) we obtain the following quadratic equation,

. 0 ) ( ) ( 2 22 13 31 40 13 22 31 2 31 22 40 2 m m m b m m m m m m m b (12) Next, we have:

...

( ) 4( )( 2). 22 13 31 2 31 22 40 2 40 13 22 31m m m m m m m m m m muv        

If Δ > 0 then there are the following two distinct roots:

, ) ( 2 ) )( ( 4 ) ( ˆ 40 22 2 31 13 31 2 22 22 40 2 31 2 40 31 22 13 40 13 31 22 1 m m m m m m m m m m m m m m m m m b         (13) . ) ( 2 ) )( ( 4 ) ( ˆ 40 22 2 31 13 31 2 22 22 40 2 31 2 40 31 22 13 40 13 31 22 2 m m m m m m m m m m m m m m m m m b         (14)

Substituting the results into the equation (12), we get: . ˆ ˆ ˆ 40 1 31 31 1 22 1 m b m m b m a   

Using the well-known Vieta’s formulas, it can be shown that a  : ˆ b1 ˆ2 ), ˆ ( / ˆ ˆ ˆ 40 1 31 2 40 1 31 31 1 22 b m bm m b m m b m   ), ˆ ( ˆ ˆ1 31 2 31 1 40 22 bm b m bm m     , ˆ ˆ ˆ ˆ1 31 2 31 2 1 40 22 bm b m bbm m     . ) ˆ ˆ ( ˆ ˆ 31 2 1 40 1 2 22 bbm b b m m    

Let us introduce notation:

 

2

31 22 40m m m m A Auv   , BB

 

muvm31m22m13m40,

 

2 22 13 31m m m m C Cuv   .

Hence, from Viete’s formula we have:

(12) Next, we have: 4 ) /( 0 ) ( ) ( ) (ba m22 b2a2 m31abba m40  ba  40 31 31 22 bm m bm m a    . (11) (12)

After putting the right site of equation (11) into the first equation of system (10) we obtain the following quadratic equation,

. 0 ) ( ) ( 2 22 13 31 40 13 22 31 2 31 22 40 2 m m m b m m m m m m m b (12) Next, we have: ). )( ( 4 ) ( 2 22 13 31 2 31 22 40 2 40 13 22 31m m m m m m m m m m      

If Δ > 0 then there are the following two distinct roots:

, ) ( 2 ˆ 40 22 2 31 40 13 31 22 1 m m m m m m m b      (13) . ) ( 2 ˆ 40 22 2 31 40 13 31 22 2 m mm mmmm b      (14)

Substituting the results into the equation (12), we get: . ˆ ˆ ˆ 40 1 31 31 1 22 1 m b m m b m a   

Using the well-known Vieta’s formulas, it can be shown that a  : ˆ1 bˆ2 ), ˆ ( / ˆ ˆ ˆ 40 1 31 2 40 1 31 31 1 22 b m bm m b m m b m   ), ˆ ( ˆ ˆ1 31 2 31 1 40 22 bm b m bm m     , ˆ ˆ ˆ ˆ1 31 2 31 2 1 40 22 bm b m bbm m     . ) ˆ ˆ ( ˆ ˆ2 1 40 1 2 31 22 bbm b b m m    

Let us introduce notation: 2

31 22

40m m

m

A  , Bm31m22m13m40, Cm31m13m222 . Hence, from Viete’s formula we have:

If Δ > 0 then there are the following two distinct roots:

4 ) /( 0 ) ( ) ( ) (ba m22 b2a2 m31abba m40  ba  40 31 31 22 bm m bm m a    . (11) (12)

After putting the right site of equation (11) into the first equation of system (10) we obtain the following quadratic equation,

. 0 ) ( ) ( 2 22 13 31 40 13 22 31 2 31 22 40 2 m m m b m m m m m m m b (12) Next, we have: ). )( ( 4 ) ( 2 22 13 31 2 31 22 40 2 40 13 22 31m m m m m m m m m m      

If Δ > 0 then there are the following two distinct roots:

, ) ( 2 ˆ 40 22 2 31 40 13 31 22 1 m mm mmmm b      (13) . ) ( 2 ˆ 40 22 2 31 40 13 31 22 2 m mm mmmm b      (14)

Substituting the results into the equation (12), we get: . ˆ ˆ ˆ 40 1 31 31 1 22 1 m b m m b m a   

Using the well-known Vieta’s formulas, it can be shown that a  : ˆ1 bˆ2 ), ˆ ( / ˆ ˆ ˆ 40 1 31 2 40 1 31 31 1 22 b m bm m b m m b m   ), ˆ ( ˆ ˆ1 31 2 31 1 40 22 bm b m bm m     , ˆ ˆ ˆ ˆ1 31 2 31 2 1 40 22 bm b m bbm m     . ) ˆ ˆ ( ˆ ˆ2 1 40 1 2 31 22 bbm b b m m    

Let us introduce notation: 2

31 22

40m m

m

A  , Bm31m22m13m40, Cm31m13m222 . Hence, from Viete’s formula we have:

(13) 4 ) /( 0 ) ( ) ( ) (ba m22 b2a2 m31abba m40  ba  40 31 31 22 bm m bm m a    . (11) (12)

After putting the right site of equation (11) into the first equation of system (10) we obtain the following quadratic equation,

. 0 ) ( ) ( 2 22 13 31 40 13 22 31 2 31 22 40 2 m m m b m m m m m m m b (12) Next, we have: ). )( ( 4 ) ( 2 22 13 31 2 31 22 40 2 40 13 22 31m m m m m m m m m m      

If Δ > 0 then there are the following two distinct roots:

, ) ( 2 ˆ 40 22 2 31 40 13 31 22 1 m m m m m m m b      (13) . ) ( 2 ˆ 40 22 2 31 40 13 31 22 2 m mm mmmm b      (14)

Substituting the results into the equation (12), we get: . ˆ ˆ ˆ 40 1 31 31 1 22 1 m b m m b m a   

Using the well-known Vieta’s formulas, it can be shown that a  : ˆ1 bˆ2

), ˆ ( / ˆ ˆ ˆ 40 1 31 2 40 1 31 31 1 22 b m bm m b m m b m   ), ˆ ( ˆ ˆ1 31 2 31 1 40 22 bm b m bm m     , ˆ ˆ ˆ ˆ1 31 2 31 2 1 40 22 bm b m bbm m     . ) ˆ ˆ ( ˆ ˆ2 1 40 1 2 31 22 bbm b b m m    

Let us introduce notation: 2

31 22

40m m

m

A  , Bm31m22m13m40, Cm31m13m222 . Hence, from Viete’s formula we have:

Cytaty

Powiązane dokumenty

Większość użytkowników forów dyskusyjnych wypowiada się na nich bardzo rzadko, w dużej części ogranicza się do czytania wypowiedzi innych (50,9%).. 30,9%

projekcie nowego kodeksu karnego Palestra 7/5(65),

Therefore, the course had four main learning objectives: (i) identify alternative theories in spatial justice, housing provision and management, and urban resilience; (ii) identify

(2016) to trace the perception of Dutch and Turkish vowels by the three groups of Polish learners of Dutch, French and English. An iden- tical L2 sound is produced authentically due

Konferencję zorganizował Oddział Materiałów Ogniotrwa- łych w Gliwicach Instytutu Ceramiki i Materiałów Budow- lanych wraz z Kołem Zakładowym SITPH przy Oddziale

Depending on the strength of inter-particle interactions and the critical range of the potential, the two-particle eigenstates are classified with respect to the excitations of

Bick 88,00 Centrozap 85,71 Espebepe 88,46 Howell 60,00 Leta 85,71 Mostostal Gdańsk 96,15 Oborniki WFM 80,00 Ocean 76,19 Pozmeat 58,33 Tonsil 96,15.. Do tych cza so we ba da nia

In this method, a global matrix is set up for each floating body and the mechanical coupling between the two vessels is calculated through the tension of hawser lines