Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Common principal components with applications in regression
Jönköping University, Jönköping International Business School, JIBS, Statistics.
Jönköping University, Jönköping International Business School, JIBS, Statistics.ORCID iD: 0000-0003-2120-0493
(English)Manuscript (preprint) (Other academic)
National Category
Probability Theory and Statistics
Identifiers
URN: urn:nbn:se:hj:diva-38586OAI: oai:DiVA.org:hj-38586DiVA, id: diva2:1175052
Available from: 2018-01-17 Created: 2018-01-17 Last updated: 2019-04-23
In thesis
1. Aspects of common principal components
Open this publication in new window or tab >>Aspects of common principal components
2017 (English)Licentiate thesis, comprehensive summary (Other academic)
Abstract [en]

The focus of this thesis is the common principal component (CPC) model, the generalization of principal components to several populations. Common principal components refer to a group of multidimensional datasets such that their inner products share the same eigenvectors and are therefore simultaneously diagonalized by a common decorrelator matrix. Common principal component analysis is essentially applied in the same areas and analysis as its one-population counterpart. The generalization to multiple populations comes at the cost of being more mathematically involved, and many problems in the area remains to be solved.

This thesis consists of three individual papers and an introduction chapter.In the first paper, the performance of two different estimation methods of the CPC model is compared for two real-world datasets and in a Monte Carlo simulation study. The second papers show that the orthogonal group and the Haar measure on this group plays an important role in PCA, both in single- and multi-population principal component analysis. The last paper considers using common principal component analysis as a tool for imposing restrictions on system-wise regression models. When the exogenous variables of a multi-dimensional model share common principal components, then each of the marginal models in the system is, up to their eigenvalues, identical. They henceform a class of regression models situated in between the classical seemingly unrelated regressions, where each set of explanatory variables is unique, and multivariate regression, where each marginal model shares the same common set of regressors.

Place, publisher, year, edition, pages
Jönköping: Jönköping University, Jönköping International Business School, 2017. p. 80
Series
JIBS Research Reports, ISSN 1403-0462 ; 2017-2
National Category
Probability Theory and Statistics
Identifiers
urn:nbn:se:hj:diva-38587 (URN)978-91-86345-79-2 (ISBN)
Supervisors
Available from: 2018-01-17 Created: 2018-01-17 Last updated: 2018-01-17Bibliographically approved
2. Applications of common principal components in multivariate and high-dimensional analysis
Open this publication in new window or tab >>Applications of common principal components in multivariate and high-dimensional analysis
2019 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

This thesis consists of four papers, all exploring some aspect of common principal component analysis (CPCA), the generalization of PCA to multiple groups. The basic assumption of the CPC model is that the space spanned by the eigenvectors is identical across several groups, whereas eigenvalues associated with the eigenvectors can vary. CPCA is used in essentially the same areas and applications as PCA.

The first paper compares the performance of the maximum likelihood and Krzanowski’s estimators of the CPC model for two real-world datasets and in a Monte Carlo simulation study. The simplicity and intuition of Krzanowski's estimator and the findings in this paper support and promote the use of this estimator for CPC models over the maximum likelihood estimator.

Paper number two uses CPCA as a tool for imposing restrictions on system-wise regression models. The paper contributes to the field by proposing a variety of explicit estimators, deriving their properties and identifying the appropriate amount of smoothing that should be imposed on the estimator. 

In the third paper, a generalization of the fixed effects PCA model to multiple populations in a CPC environment is proposed. The model includes mainly geometrical, rather than probabilistic, assumptions, and is designed to account for any possible prior information about the noise in the data to yield better estimates, obtained by minimizing a least squares criterion with respect to a specified metric.

The fourth paper survey some properties of the orthogonal group and the associated Haar measure on it. It is demonstrated how seemingly abstract results contribute to applied statistics and, in particular, to PCA.

Abstract [sv]

Denna avhandling består av fyra papper som alla utforskar någon del av gemensam principalkomponentanalys (CPCA), generaliseringen av principal-komponentanalys (PCA) till flera grupper. Det grundläggande antagandet av CPC-modellen är att egenvektorerna är identiska för samtliga grupper medan de associerade egenvärdena kan variera.

Det första pappret jämför prestationen av maximum likelihood estimatorn och Krzanowskis estimator för CPC-modellen för två verkliga dataset och i en Monte Carlo-simuleringstudie. Enkelheten och intuitionen av Krzanowskis estimator samt resultaten i detta papper stödjer användningen av denna estimator för CPC-modeller över maximum likelihood-estimatorn.

Papper nummer två använder CPCA som ett verktyg för att införa restriktioner på systemvisa regressionsmodeller. Pappret bidrar till området genom att föreslå en rad olika estimatorer, härleda deras egenskaper och identifiera lämplig mängd utjämning som ska åläggas estimatorn.

I det tredje pappret föreslås en generalisering av PCA-modellen med icke-stokastiska effekter till flera populationer i en CPC-miljö. Modellen innehåller huvudsakligen geometriska, snarare än probabilistiska antaganden och är utformad för att betrakta eventuell information om bruset i dataseten för att ge bättre uppskattningar; erhållna genom att minimera ett minsta kvadratkriterium med avseende på ett specificerat metriskt rum.

Det fjärde pappret undersöker egenskaper hos den ortogonala gruppen och det associerade Haar-måttet på gruppen. Det demonstreras hur till synes abstrakta resultat är viktiga för tillämpad statistik och i synnerhet för PCA.

Place, publisher, year, edition, pages
Jönköping: Jönköping University, Jönköping International Business School, 2019. p. 56
Series
JIBS Dissertation Series, ISSN 1403-0470 ; 131
National Category
Probability Theory and Statistics
Identifiers
urn:nbn:se:hj:diva-43519 (URN)978-91-86345-93-8 (ISBN)
Public defence
2019-05-17, B1014, Jönköping International Business School, Jönköping, 10:00 (English)
Opponent
Supervisors
Available from: 2019-04-23 Created: 2019-04-23 Last updated: 2019-04-23Bibliographically approved

Open Access in DiVA

No full text in DiVA

Authority records

Duras, ToniHolgersson, Thomas

Search in DiVA

By author/editor
Duras, ToniHolgersson, Thomas
By organisation
JIBS, Statistics
Probability Theory and Statistics

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 356 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf