• Nem Talált Eredményt

Modelling local GPS/levelling geoid undulations using Support Vector Machines

N/A
N/A
Protected

Academic year: 2022

Ossza meg "Modelling local GPS/levelling geoid undulations using Support Vector Machines"

Copied!
5
0
0

Teljes szövegt

(1)

Ŕ periodica polytechnica

Civil Engineering 52/1 (2008) 39–43 doi: 10.3311/pp.ci.2008-1.06 web: http://www.pp.bme.hu/ci c Periodica Polytechnica 2008 RESEARCH ARTICLE

Modelling local GPS/levelling geoid undulations using Support Vector Machines

PiroskaZaletnyik/LajosVölgyesi/BélaPaláncz

Received 2007-12-05

Abstract

Support vector machines (SVM) with wavelet kernel has been applied to the correcting gravimetric geoid using GPS/levelling data. These data were divided into a training and a validation set in order to ensure the extendability of the approximation of the corrector surface. The optimal parameters of the SVM were considered as a trade-offbetween accuracy and extendability of the solution in order to avoid overlearning. Employing 194 training points and 110 validation points, SVM provided an ap- proximation with less than 3 cm standard deviation of the error and nearly perfect extendability.

Keywords

geoid·corrector surface·GPS·support vector regression· wavelet kernel

Acknowledgement

The authors wish to thank A. Kenyeres and the Hungarian Institute of Geodesy, Cartography and Remote Sensing for pro- viding GPS/levelling data of Hungary and also the Hungarian Research Fund OTKA project T046718.

Piroska Zaletnyik

Department of Geodesy and Surveying, Research Group of Physical Geodesy and Geodynamics of the Hungarian Academy of Sciences, BME, POBox91, H- 1521, Hungary

e-mail: zaletnyikp@hotmail.com

Lajos Völgyesi

Department of Geodesy and Surveying, Research Group of Physical Geodesy and Geodynamics of the Hungarian Academy of Sciences, BME, POBox91, H- 1521, Hungary

e-mail: volgyesi@eik.bme.hu

Béla Paláncz

Department of Photogrammetry and Geoinformatics, BME, POBox91, H-1521, Hungary

e-mail: palancz@epito.bme.hu

1 Introduction

The accuracy of the gravimetrical geoid can be significantly improved using GPS/levelling measurements. The new, ad- justed geoid can be constructed as the gravimetric one plus the so called corrector surface, the difference between the gravimet- ric and the GPS/levelling geoid.

Recently, wide variety of higher-order parametric and non- parametric surfaces have been used as corrector surfaces, such as polynomial models by Fotopoulos and Sideris 2005 [4], spline interpolation by Featherstone 2000 [2] and Zaletnyik et al 2007 [13], least squares collocation (LSC) by Iliffe et al 2003 [5], kriging by Nahavandchi and Soltanpour 2004 [8], combined least squares adjustments by Fotopoulos 2005 [3], and various other surfaces. Most recently Zaletnyik et al. 2007 [14] em- ployed thin plate spline (TPS) surface, solving the problem via finite element method. Suffice it to say, there are numerous surface-fitting options, each with their own advantages and dis- advantages, which will not be discussed nor debated here.

Concerning application of soft computing technique Kav- zoglu and Saka 2005 [6] and Lao-Sheng Lin 2006 [7] em- ployed artificial neural network (ANN) for approximating the GPS/levelling geoid instead of the corrector surface itself. Both of them applied feed-forward ANN with the standard sigmoid activation functions and different number of hidden layers. Za- letnyik et al. 2007 [13] also used ANN but with radial bases activation function (RBF) and regularization in the training phase. Soltanpour et al 2006 [11] used second generation wavelets to approximate corrector surface directly. This tech- nique let extend the classical wavelet approximation, which re- quires regularly spaced/sampled data, for unregularly spaced dataset, Sweldens 1997 [12].

Another soft computing technique is represented by the sup- port vector machines (SVM), which are learning algorithms that have many applications in pattern recognition and nonlinear re- gression. In this study we propose to apply support vector ma- chine with wavelet kernel for modelling the corrector surface.

(2)

2 Support Vector Machines for Regression

The problem of regression is that of finding a function which approximates mapping from an input domain to the real num- bers based on a training sample. We refer to the difference be- tween the hypothesis output and its training value as the residual of the output, an indication of the accuracy of the fit at this point.

We must decide how to measure the importance of this accuracy, as small residuals may be inevitable while we wish to avoid large ones. The loss function determines this measure. Each choice of loss function will result in a different overall strategy for per- forming regression. For example least square regression uses the sum of the squares of the residuals.

Although several different approaches are possible, we will provide an analysis for generalization of regression by introduc- ing a threshold test accuracy, beyond which we consider a mis- take to have been made. We therefore aim to provide a bound on the probability that a randomly drawn validation point will have accuracy less than∈. One way of visualizing this method of as- sessing performance is to consider a band of size± ∈around the hypothesis function any training points lying outside this band are considered to be training mistakes, see Fig. 1.

Fig. 1. Linear∈-insensitive loss functionL(x,y,f)

Therefore we can define a so called∈-insensitive loss func- tion. The linear∈-insensitive loss functionL(x,y, f)is de- fined by

L(x,y, f)=(|y− f(x)|)=max(0, |y− f(x)| − ∈) (1) where f is a real-valued function on a domainX ⊂ <n,x∈ X andy ∈ <. Similarly the quadratic∈-insensitive loss is given by

L2(x,y, f)=(|y− f(x)|)2. (2) Support vector regression (SVR) uses an admissible kernel, which satisfies the Mercer’s condition to map the data in in- put space to a highdimensional feature space in which we can process a regression problem in linear form. Letx ∈ <n and y ∈ <, where <n represents input space, see Cristianini and Shawe-Taylor 2003 [1]. By some nonlinear mapping8, the vec- torxis mapped into a feature space in which a linear regressor function is defined,

y= f(x, w)= hw, 8(x)i +b. (3)

We seek to estimate this f function based on independent uni- formly distributed data {{x1,y1}, ...,{xm,ym}}, by finding w which minimizing the quadratic ∈-insensitive gosses, with ∈, namely the following function should be minimize

c

m

X

i=1

L2(xi,yi,f)+1

2(kwk)2→min (4) wherewis weight vector andcis a dimensionless constant pa- rameter. Considering dual representation of a linear regressor in (3), f(x)can be expressed as

f(x)=

m

X

i=1

βiyih8(xi), 8(x)i +b (5) what means that the regressor can be expressed as a linear com- bination of the training points. Consequently using an admissi- ble kernel, a kernel satisfying the Mercer’s condition, Paláncz et al 2005 [10], we get

f(x)=

m

X

i=1

βiyiK(xi, x)+b=

m

X

i=1

αiK(xi, x)+b. (6) By using Lagrange multiplier techniques, the minimization problem of (4) leads to the following dual optimization problem

maximize W(α)=

m

P

i=1

yiαi− ∈

m

P

i=1

i|

12

m

P

i,j=1

αiαj

K(xi, xj)+1cδi j

(7)

subject to

m

P

i=1

αi =0.

Let

f(x)=

m

X

i=1

αiK(xi, x)+b, (8) whereα is the solution of the quadratic optimization problem andbis chosen so that f(xi)=yi− ∈ −αci for anyαi >0.

For samples are inside the∈-tube,{xi : |f(xi)−yi|<∈}, the correspondingαis zero. It means we do not need these samples to describe the weight vectorw.Consequently

f(x)= X

iSV

αiK(xi, x)+b (9) where

SV = {i: |f(xi)−yi| ≥ ∈}. (10) Thesexi sample vectors,{xi : i ∈SV}, that come with nonva- nishing coefficientαare called support vectors.

3 Wavelet Kernel

In our case, we select wavelet kernel forn = 2, which pro- vides better local characterization than other kernels, see Zhang et al. 2004 [15] and was proved to be very efficient in many regression problems, e.g. Paláncz et al. 2005 [10].

Wavelet kernel witha ∈ <1and all compactX ⊂ <n, K(x,z)=

n

Y

i=1

cos

1.75xi −zi a

exp

"

−(xi −zi)2 2a2

#!

. (11)

(3)

4 Dataset for the numerical computations

The original gravimetric geoid was modelled via third order spline approximation, which provides a fairly good approxima- tion, a fitting with 1 - 2 cm error in height, see Zaletnyik et al.

2007 [13].

Fig. 2. The Hungarian gravimetric geoid

For modelling the corrector surface, there are 304 GPS/levelling data available. One part of these data was em- ployed as training set (194 measurements), the other part was used as validation set (110 measurement), see Fig. 3.

Fig. 3. GPS/levelling data - the training set (circles), and the validation set (triangles)

5 Parameter Study

In order to achieve an efficient fitting, one should find the op- timal parameter of the applied kernel function (a)as well as the proper values ofcand∈. Parameter investigations showed that with increasing values ofcand∈, the regression error (root mean square error,RMSE) decreases on the training and the val- idation set, too. In our casec=400 and∈= 103 proved to be reasonable values, while for greater values the changes of RMSE’s are negligable.

However the value of the parametera has a strong influence on the quality of the approximation. Table??shows the change

of the sum ofRMSE’s (that of training and validation set, respec- tively) as well as the ratio of these RMSE’s, namely introducing

η= R M S EV

R M S Et (12)

a ratio indicates how realiably can we extend our regression model for not measured data. The ideal value is 1. Ifη >>

1, then so-called overlearning effect takes place.

Tab. 1. The result of the parameter study in case ofc=400 and∈=10−3

a R M S Et[cm] R M S EV [cm] R M S Etotal[cm] η

0.50 0.50 2.06 1.28 4.12

1.00 1.64 1.94 1.79 1.18

1.50 2.15 2.37 2.26 1.10

2.00 2.31 2.57 2.44 1.11

2.50 2.59 2.74 2.66 1.06

Fig. 4 shows the corrector surface in case ofa=0.5 when the total error is smallR M S Etotal(=1.28) butη=4.12 is high.

Fig. 4.Corrector surface in case of typical overlearning (a=0.5)

This result indicates that one should make a trade-off be- tween extendability (η) and regression error (R M S Etotal). In our case we selecteda=2.5, which ensures smallηand accept- ableRMSEas well as smooth regression surface, see Fig. 5.

6 The model for the corrector surface

Using these parameter values (c=400 and∈= 103m and a=2.5) the computation with the wavelet kernel Eq. (11) was carried out. We used theMathematicaimplementation of SVM regression, see Paláncz 2005 [9].

(4)

Tab. 2. Corrector Surface approximated by SVM

Training set Validation set

Method SD [cm] Min [cm] Max [cm] R M S EV[cm] SD [cm] Min [cm] Max [cm] R M S EV [cm] η

SVM regression 2.60 -7.74 6.47 2.59 2.75 -7.68 5.59 2.74 1.06

Fig. 5. Smooth and extendable corrector surface

The analytical form of the corrector surface is 1H = −0.0370989

−8.52042e0.08(47.2429−ϕ)20.08(16.4483−λ)2

·cos[0.7(47.2429−ϕ)] cos[0.7(16.4483−λ)]

−11.4187e0.08(47.0987−ϕ)20.08(16.5562−λ)2

·cos[0.7(47.0987−ϕ)] cos[0.7(16.5562−λ)] +...

−6.64333e0.08(48.0101−ϕ)20.08(22.5098−λ)2

·cos[0.7(48.0101−ϕ)] cos[0.7(22.5098−λ)]

−4.17322e0.08(48.1296−ϕ)20.08(22.5491−λ)2

·cos[0.7(48.1296−ϕ)] cos[0.7(22.5491−λ)]

−1.67954e0.08(47.9176−ϕ)20.08(22.762−λ)2

·cos[0.7(47.9176−ϕ)] cos[0.7(22.762−λ)]. (13) Table 1 shows that the standard deviation (SD) on the training as well as on the validation set is about 2.6 - 2.8 cm, which from practical point of view is acceptable and which is also important, that extendability coefficient is very good, near to unity. In the table RMSE- root mean squared error - is the square root of the mean of the error vector of the measurement points,

R M S E= s

MeanX

i

(1H(ϕi, λi)−1Hi)2 (14)

The resulted corrector surface now is very smooth, see Fig. 5.

7 Adjusted Geoid

In order to get the corrected geoid, the corrector surface should be added to the original geoid surface, see Fig. 6.

Fig. 6. Adjusted geoid with the training and validation points

8 Conclusions

SVM with quadratic∈-insensitive loss function was applied to constructing corrector surface for gravimetrical geoid, using GPS/levelling data. Employing wavelet kernel it turned out, that only the kernel parametera has considerable influence on the quality of the approximation, while the SVM parameterscand

∈do not play important role in this case. The optimal parame- ters of the SVM were considered as a trade-offbetween accuracy and extendability of the solution in order to avoid overlearning.

Employing 194 training points and 110 validation points, SVM provided an approximation with less than 3 cm standard devia- tion of the error and nearly perfect extendability. The corrector surface can be described via analytical form and directly im- plemented in a high level language, like C, in order to get high performance evaluation.

In the future, the investigation of the application of other type of kernel can be reasonable.

References

1 Cristianini N, Shawe-Taylor J,An introduction to Support Vector Machines and other kernel - based learning methods, University Press, Cambridge, 2003.

2 Featherstone WE,Refinement of a gravimetric geoid using GPS and level- ling data, J. Surv. Eng.126(2000), no. 2, 27-56.

3 Fotopoulos G,Calibration of geoid error models via a combined adjust- ment of ellipsoidal, orthometric and gravimetric geoid height data, J. of Geod.79(2005), no. 1-3, 111-123.

4 Fotopoulos G, Sideris MG,Spatial Modelling and Analysis of Adjusted Residuals over a network of GPS-levelling Benchmarks, GEOMATICA59 (2005), no. 3, 251-262.

(5)

5 Iliffe JC, Ziebart M, Cross PA, Forsberg R, Strykowski G, Tschern- ing CC,OGSM02: a new model for converting GPS-derived heights to local height datums in Great Britain and Ireland, Surv. Rev.37(2003), no. 290, 276-293.

6 Kavzoglu T, Saka MH,Modelling local GPS/levelling geoid undulations using artificial neural networks, J. Geod.78(2005), 520-527.

7 Lao-Sheng Lin,Application of a Back-Propagation Artificial Neural Net- work to Regional Grid-Based Geoid Model Generation Using GPS and Lev- elling Data, J. of Surv. Eng.133(2007), no. 2, 81-89.

8 Nahavandchi H, Soltanpour A,An attempt to define a new height datum in Norvay. The geodesy and Hydrography days 2004, 4-5 Nov. Sandnes, Nor- way.

9 Paláncz B,Support Vector Regression via Mathematica(2005), available at http://library.wolfram.com/infocenter/Mathsource/5270/.

10Paláncz B, Völgyesi L, Popper Gy,Support Vector Regression via Mathe- matica, Periodica Polytechnica, Civil Eng49(2005), no. 1, 57-84.

11Soltanpour A, Nahavandchi H, Featherstone WE,Geoid-type surface determination using wavelet-based combination of gravimetric quasi/geoid and GPS/levelling data, Geophysical Research Abstract8(2006), 04612.

12Sweldens W, The lifting scheme: a construction of second generation wavelets, SIAM J Math. Anal29(1997), no. 2, 511-546.

13Zaletnyik P, Paláncz B, Völgyesi L, Kenyeres A,Correction of the gravi- metric geoid using GPS/levelling data, Geomatikai KözleményekX(2007), 231-240 (Hungarian).

14Zaletnyik P, Völgyesi L, Kirchner I, Paláncz B, Combination of GPS/levelling and gravimetric geoid by using the thin plate spline interpola- tion technique via finite element method, J. of Appl. Geod1(2007), 233-239.

issue 4.

15Zhang L, Zhou W, Jiao L,Wavelet Support Vector Machine, IEEE Trans.

Systems, Man and Cybernetics - Part B: Cybernetics, 2004, pp. 34-39.

Hivatkozások

KAPCSOLÓDÓ DOKUMENTUMOK

(2020) &#34;Lane Change Prediction Using Gaussian Classification, Support Vector Classification and Neural Network Classifiers&#34;, Periodica Polytechnica Transportation

Only effective parameters (inde- pendent variables) including RQD, UCS, maximum depth of hole, blast hole coupling ratio and tunnel cross section area were considered in the

To reach the highest accuracy, geodesists use GPS for relative coordinate- determination. geoid heights at the two relative points are nearly the same), then we can use

smart machines, Precision Agriculture, GPS, data analysing, Vibration

The Stack Overflow based experiments applied Multinomial, Gaussian- and Bernoulli Naive Bayes, Support Vector Machine with linear kernel, Linear Logistic Regression, Decision

The composite stable isotope records of the Baradla Cave speleothem were compared with tem- perature and precipitation-amount data at a monthly resolu- tion using wavelet

A tanulmány a support vector gépek (SVM) alkalmazásának lehetőségét mutatja be hitelpontozó kártyák fejlesztésére.. A módszer segítségével egy publikusan elérhető

An enormous acoustic mismatch between normally phonated (neutral) and whispered speech is the main reason why modern Automatic Speech Recognition (ASR) systems