IBM SPSS Categories 21

Total Page:16

File Type:pdf, Size:1020Kb

IBM SPSS Categories 21 i IBM SPSS Categories 21 Jacqueline J. Meulman Willem J. Heiser Note: Before using this information and the product it supports, read the general information under Notices on p. 301. This edition applies to IBM® SPSS® Statistics 21 and to all subsequent releases and modifications until otherwise indicated in new editions. Adobe product screenshot(s) reprinted with permission from Adobe Systems Incorporated. Microsoft product screenshot(s) reprinted with permission from Microsoft Corporation. Licensed Materials - Property of IBM © Copyright IBM Corporation 1989, 2012. U.S. Government Users Restricted Rights - Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Preface IBM® SPSS® Statistics is a comprehensive system for analyzing data. The Categories optional add-on module provides the additional analytic techniques described in this manual. The Categories add-on module must be used with the SPSS Statistics Core system and is completely integrated into that system. About IBM Business Analytics IBM Business Analytics software delivers complete, consistent and accurate information that decision-makers trust to improve business performance. A comprehensive portfolio of business intelligence, predictive analytics, financial performance and strategy management,andanalytic applications provides clear, immediate and actionable insights into current performance and the ability to predict future outcomes. Combined with rich industry solutions, proven practices and professional services, organizations of every size can drive the highest productivity, confidently automate decisions and deliver better results. As part of this portfolio, IBM SPSS Predictive Analytics software helps organizations predict future events and proactively act upon that insight to drive better business outcomes. Commercial, government and academic customers worldwide rely on IBM SPSS technology as a competitive advantage in attracting, retaining and growing customers, while reducing fraud and mitigating risk. By incorporating IBM SPSS software into their daily operations, organizations become predictive enterprises – able to direct and automate decisions to meet business goals and achieve measurable competitive advantage. For further information or to reach a representative visit http://www.ibm.com/spss. Technical support Technical support is available to maintenance customers. Customers may contact Technical Support for assistance in using IBM Corp. products or for installation help for one of the supported hardware environments. To reach Technical Support, see the IBM Corp. web site at http://www.ibm.com/support. Be prepared to identify yourself, your organization, and your support agreement when requesting assistance. Technical Support for Students If you’re a student using a student, academic or grad pack version of any IBM SPSS software product, please see our special online Solutions for Education (http://www.ibm.com/spss/rd/students/) pages for students. If you’re a student using a university-supplied copy of the IBM SPSS software, please contact the IBM SPSS product coordinator at your university. Customer Service If you have any questions concerning your shipment or account, contact your local office. Please have your serial number ready for identification. © Copyright IBM Corporation 1989, 2012. iii Training Seminars IBM Corp. provides both public and onsite training seminars. All seminars feature hands-on workshops. Seminars will be offered in major cities on a regular basis. For more information on these seminars, go to http://www.ibm.com/software/analytics/spss/training. Acknowledgements The optimal scaling procedures and their implementation in IBM® SPSS® Statistics were developed by the Data Theory Scaling System Group (DTSS), consisting of members of the departments of Education and Psychology of the Faculty of Social and Behavioral Sciences at Leiden University. Willem Heiser, Jacqueline Meulman, Gerda van den Berg, and Patrick Groenen were involved with the original 1990 procedures. Jacqueline Meulman and Peter Neufeglise participated in the development of procedures for categorical regression, correspondence analysis, categorical principal components analysis, and multidimensional scaling. In addition, Anita van der Kooij contributed especially to CATREG, CORRESPONDENCE, and CATPCA. Willem Heiser, Jacques Commandeur, Frank Busing, Gerda van den Berg, and Patrick Groenen participated in the development of the PROXSCAL procedure. Frank Busing, Willem Heiser, Patrick Groenen, and Peter Neufeglise participated in the development of the PREFSCAL procedure. iv Contents Part I: User's Guide 1 Introduction to Optimal Scaling Procedures for Categorical Data 1 WhatIsOptimalScaling?....................................................... 1 WhyUseOptimalScaling?...................................................... 1 OptimalScalingLevelandMeasurementLevel ...................................... 2 SelectingtheOptimalScalingLevel............................................ 3 TransformationPlots....................................................... 3 CategoryCodes .......................................................... 4 WhichProcedureIsBestforYourApplication?...................................... 6 CategoricalRegression..................................................... 7 CategoricalPrincipalComponentsAnalysis ..................................... 7 NonlinearCanonicalCorrelationAnalysis....................................... 8 CorrespondenceAnalysis................................................... 9 MultipleCorrespondenceAnalysis............................................10 MultidimensionalScaling....................................................11 MultidimensionalUnfolding..................................................11 AspectRatioinOptimalScalingCharts ............................................12 RecommendedReadings.......................................................12 2 Categorical Regression (CATREG) 14 DefineScaleinCategoricalRegression............................................15 CategoricalRegressionDiscretization.............................................16 CategoricalRegressionMissingValues............................................17 CategoricalRegressionOptions..................................................18 CategoricalRegressionRegularization.............................................20 CategoricalRegressionOutput...................................................21 CategoricalRegressionSave....................................................23 CategoricalRegressionTransformationPlots........................................24 CATREGCommandAdditionalFeatures.............................................25 © Copyright IBM Corporation 1989, 2012. v 3 Categorical Principal Components Analysis (CATPCA) 26 DefineScaleandWeightinCATPCA...............................................28 CategoricalPrincipalComponentsAnalysisDiscretization..............................30 CategoricalPrincipalComponentsAnalysisMissingValues.............................31 CategoricalPrincipalComponentsAnalysisOptions...................................32 CategoricalPrincipalComponentsAnalysisOutput....................................34 CategoricalPrincipalComponentsAnalysisSave.....................................35 CategoricalPrincipalComponentsAnalysisObjectPlots................................36 CategoricalPrincipalComponentsAnalysisCategoryPlots..............................37 CategoricalPrincipalComponentsAnalysisLoadingPlots..............................38 CATPCACommandAdditionalFeatures.............................................39 4 Nonlinear Canonical Correlation Analysis (OVERALS) 40 DefineRangeandScale........................................................43 DefineRange................................................................43 NonlinearCanonicalCorrelationAnalysisOptions....................................44 OVERALSCommandAdditionalFeatures............................................45 5 Correspondence Analysis 46 DefineRowRangeinCorrespondenceAnalysis......................................47 DefineColumnRangeinCorrespondenceAnalysis....................................48 CorrespondenceAnalysisModel.................................................49 CorrespondenceAnalysisStatistics...............................................51 CorrespondenceAnalysisPlots..................................................52 CORRESPONDENCECommandAdditionalFeatures....................................54 6 Multiple Correspondence Analysis 55 DefineVariableWeightinMultipleCorrespondenceAnalysis............................57 MultipleCorrespondenceAnalysisDiscretization.....................................57 MultipleCorrespondenceAnalysisMissingValues....................................58 MultipleCorrespondenceAnalysisOptions..........................................60 vi MultipleCorrespondenceAnalysisOutput..........................................61 MultipleCorrespondenceAnalysisSave............................................63 MultipleCorrespondenceAnalysisObjectPlots......................................63 MultipleCorrespondenceAnalysisVariablePlots.....................................64 MULTIPLECORRESPONDENCECommandAdditionalFeatures...........................65 7 Multidimensional Scaling (PROXSCAL) 67 ProximitiesinMatricesacrossColumns............................................69 ProximitiesinColumns.........................................................70 ProximitiesinOneColumn......................................................71
Recommended publications
  • Canonical Correlation a Tutorial
    Canonical Correlation a Tutorial Magnus Borga January 12, 2001 Contents 1 About this tutorial 1 2 Introduction 2 3 Definition 2 4 Calculating canonical correlations 3 5 Relating topics 3 5.1 The difference between CCA and ordinary correlation analysis . 3 5.2Relationtomutualinformation................... 4 5.3 Relation to other linear subspace methods ............. 4 5.4RelationtoSNR........................... 5 5.4.1 Equalnoiseenergies.................... 5 5.4.2 Correlation between a signal and the corrupted signal . 6 A Explanations 6 A.1Anoteoncorrelationandcovariancematrices........... 6 A.2Affinetransformations....................... 6 A.3Apieceofinformationtheory.................... 7 A.4 Principal component analysis . ................... 9 A.5 Partial least squares ......................... 9 A.6 Multivariate linear regression . ................... 9 A.7Signaltonoiseratio......................... 10 1 About this tutorial This is a printable version of a tutorial in HTML format. The tutorial may be modified at any time as will this version. The latest version of this tutorial is available at http://people.imt.liu.se/˜magnus/cca/. 1 2 Introduction Canonical correlation analysis (CCA) is a way of measuring the linear relationship between two multidimensional variables. It finds two bases, one for each variable, that are optimal with respect to correlations and, at the same time, it finds the corresponding correlations. In other words, it finds the two bases in which the correlation matrix between the variables is diagonal and the correlations on the diagonal are maximized. The dimensionality of these new bases is equal to or less than the smallest dimensionality of the two variables. An important property of canonical correlations is that they are invariant with respect to affine transformations of the variables. This is the most important differ- ence between CCA and ordinary correlation analysis which highly depend on the basis in which the variables are described.
    [Show full text]
  • A Tutorial on Canonical Correlation Analysis
    Finding the needle in high-dimensional haystack: A tutorial on canonical correlation analysis Hao-Ting Wang1,*, Jonathan Smallwood1, Janaina Mourao-Miranda2, Cedric Huchuan Xia3, Theodore D. Satterthwaite3, Danielle S. Bassett4,5,6,7, Danilo Bzdok,8,9,10,* 1Department of Psychology, University of York, Heslington, York, YO10 5DD, United Kingdome 2Centre for Medical Image Computing, Department of Computer Science, University College London, London, United Kingdom; MaX Planck University College London Centre for Computational Psychiatry and Ageing Research, University College London, London, United Kingdom 3Department of Psychiatry, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, 19104, USA 4Department of Bioengineering, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA 5Department of Electrical and Systems Engineering, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA 6Department of Neurology, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104 USA 7Department of Physics & Astronomy, School of Arts & Sciences, University of Pennsylvania, Philadelphia, PA 19104 USA 8 Department of Psychiatry, Psychotherapy and Psychosomatics, RWTH Aachen University, Germany 9JARA-BRAIN, Jülich-Aachen Research Alliance, Germany 10Parietal Team, INRIA, Neurospin, Bat 145, CEA Saclay, 91191, Gif-sur-Yvette, France * Correspondence should be addressed to: Hao-Ting Wang [email protected] Department of Psychology, University of York, Heslington, York, YO10 5DD, United Kingdome Danilo Bzdok [email protected] Department of Psychiatry, Psychotherapy and Psychosomatics, RWTH Aachen University, Germany 1 1 ABSTRACT Since the beginning of the 21st century, the size, breadth, and granularity of data in biology and medicine has grown rapidly. In the eXample of neuroscience, studies with thousands of subjects are becoming more common, which provide eXtensive phenotyping on the behavioral, neural, and genomic level with hundreds of variables.
    [Show full text]
  • A Call for Conducting Multivariate Mixed Analyses
    Journal of Educational Issues ISSN 2377-2263 2016, Vol. 2, No. 2 A Call for Conducting Multivariate Mixed Analyses Anthony J. Onwuegbuzie (Corresponding author) Department of Educational Leadership, Box 2119 Sam Houston State University, Huntsville, Texas 77341-2119, USA E-mail: [email protected] Received: April 13, 2016 Accepted: May 25, 2016 Published: June 9, 2016 doi:10.5296/jei.v2i2.9316 URL: http://dx.doi.org/10.5296/jei.v2i2.9316 Abstract Several authors have written methodological works that provide an introductory- and/or intermediate-level guide to conducting mixed analyses. Although these works have been useful for beginning and emergent mixed researchers, with very few exceptions, works are lacking that describe and illustrate advanced-level mixed analysis approaches. Thus, the purpose of this article is to introduce the concept of multivariate mixed analyses, which represents a complex form of advanced mixed analyses. These analyses characterize a class of mixed analyses wherein at least one of the quantitative analyses and at least one of the qualitative analyses both involve the simultaneous analysis of multiple variables. The notion of multivariate mixed analyses previously has not been discussed in the literature, illustrating the significance and innovation of the article. Keywords: Mixed methods data analysis, Mixed analysis, Mixed analyses, Advanced mixed analyses, Multivariate mixed analyses 1. Crossover Nature of Mixed Analyses At least 13 decision criteria are available to researchers during the data analysis stage of mixed research studies (Onwuegbuzie & Combs, 2010). Of these criteria, the criterion that is the most underdeveloped is the crossover nature of mixed analyses. Yet, this form of mixed analysis represents a pivotal decision because it determines the level of integration and complexity of quantitative and qualitative analyses in mixed research studies.
    [Show full text]
  • A Probabilistic Interpretation of Canonical Correlation Analysis
    A Probabilistic Interpretation of Canonical Correlation Analysis Francis R. Bach Michael I. Jordan Computer Science Division Computer Science Division University of California and Department of Statistics Berkeley, CA 94114, USA University of California [email protected] Berkeley, CA 94114, USA [email protected] April 21, 2005 Technical Report 688 Department of Statistics University of California, Berkeley Abstract We give a probabilistic interpretation of canonical correlation (CCA) analysis as a latent variable model for two Gaussian random vectors. Our interpretation is similar to the proba- bilistic interpretation of principal component analysis (Tipping and Bishop, 1999, Roweis, 1998). In addition, we cast Fisher linear discriminant analysis (LDA) within the CCA framework. 1 Introduction Data analysis tools such as principal component analysis (PCA), linear discriminant analysis (LDA) and canonical correlation analysis (CCA) are widely used for purposes such as dimensionality re- duction or visualization (Hotelling, 1936, Anderson, 1984, Hastie et al., 2001). In this paper, we provide a probabilistic interpretation of CCA and LDA. Such a probabilistic interpretation deepens the understanding of CCA and LDA as model-based methods, enables the use of local CCA mod- els as components of a larger probabilistic model, and suggests generalizations to members of the exponential family other than the Gaussian distribution. In Section 2, we review the probabilistic interpretation of PCA, while in Section 3, we present the probabilistic interpretation of CCA and LDA, with proofs presented in Section 4. In Section 5, we provide a CCA-based probabilistic interpretation of LDA. 1 z x Figure 1: Graphical model for factor analysis. 2 Review: probabilistic interpretation of PCA Tipping and Bishop (1999) have shown that PCA can be seen as the maximum likelihood solution of a factor analysis model with isotropic covariance matrix.
    [Show full text]
  • Canonical Correlation Analysis and Graphical Modeling for Huaman
    Canonical Autocorrelation Analysis and Graphical Modeling for Human Trafficking Characterization Qicong Chen Maria De Arteaga Carnegie Mellon University Carnegie Mellon University Pittsburgh, PA 15213 Pittsburgh, PA 15213 [email protected] [email protected] William Herlands Carnegie Mellon University Pittsburgh, PA 15213 [email protected] Abstract The present research characterizes online prostitution advertisements by human trafficking rings to extract and quantify patterns that describe their online oper- ations. We approach this descriptive analytics task from two perspectives. One, we develop an extension to Sparse Canonical Correlation Analysis that identifies autocorrelations within a single set of variables. This technique, which we call Canonical Autocorrelation Analysis, detects features of the human trafficking ad- vertisements which are highly correlated within a particular trafficking ring. Two, we use a variant of supervised latent Dirichlet allocation to learn topic models over the trafficking rings. The relationship of the topics over multiple rings character- izes general behaviors of human traffickers as well as behaviours of individual rings. 1 Introduction Currently, the United Nations Office on Drugs and Crime estimates there are 2.5 million victims of human trafficking in the world, with 79% of them suffering sexual exploitation. Increasingly, traffickers use the Internet to advertise their victims’ services and establish contact with clients. The present reserach chacterizes online advertisements by particular human traffickers (or trafficking rings) to develop a quantitative descriptions of their online operations. As such, prediction is not the main objective. This is a task of descriptive analytics, where the objective is to extract and quantify patterns that are difficult for humans to find.
    [Show full text]
  • SOA Education Day for CMU Students and Faculty
    IBM SOA Roadshow SOA Education Day For CMU Students and Faculty Frank Stein Program Director IBM Federal SOA Institute © 2007 IBM Corporation IBM SOA Architect Summit Agenda 9:00 – 9:15am Welcome – Dr. Randy Bryant, Dean, SCS 9:15 – 9:45am Introduction to SOA – Frank Stein 9:45 – 10:15am Leveraging Information in SOA – Mark Sherman 10:15 – 10:45am SOA Governance and SOA Lifecycle – Naveen Sachdeva 10:45 – 11:15am Reuse, Community, Ecosystems & Catalogs – Jay Palat 11:15 – 12:00pm IBM’s SOA Experience, Trends, Outlook – Sanjay Bose 12:00 – 1:15pm Lunch Break & Faculty Reception 1:15 – 2:00pm Business Process Modeling Overview – Tom McManus & Emilio Zegarra 2:00 – 3:15pm Innov8 3D BPM “Serious Game” – Dave Daniel 2 SOA Education Day at CMU IBM SOA Roadshow Overview Presentation: SOA: A Foundation for Improved Agility and Economic Value Frank Stein Program Director IBM Federal SOA Institute © 2007 IBM Corporation IBM SOA Architect Summit Cloud Computing Announcement – Oct 10, 2007 • IBM and Google will provide hw, sw, and services to assist the academic community to explore Internet-scale computing. •CMU, MIT, Stanford, U of Washington, UC-Berkeley, U of MD 4 SOA Education Day at CMU IBM SOA Architect Summit Agenda Why SOA SOA Technical Concepts What’s Next for SOA Summary 5 SOA Education Day at CMU IBM SOA Architect Summit Businesses are Placing a Premium on Innovation THINK THINK 6 SOA Education Day at CMU IBM SOA Architect Summit Innovation Impacts Business Models Is Your Architecture Ready? “ On a flat earth, the most important
    [Show full text]
  • Cross-Modal Subspace Clustering Via Deep Canonical Correlation Analysis
    The Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI-20) Cross-Modal Subspace Clustering via Deep Canonical Correlation Analysis Quanxue Gao,1 Huanhuan Lian,1∗ Qianqian Wang,1† Gan Sun2 1State Key Laboratory of Integrated Services Networks, Xidian University, Xi’an 710071, China. 2State Key Laboratory of Robotics, Shenyang Institute of Automation, Chinese Academy of Sciences, China. [email protected], [email protected], [email protected], [email protected] Abstract Different Different Self- S modalities subspaces expression 1 For cross-modal subspace clustering, the key point is how to ZZS exploit the correlation information between cross-modal data. Deep 111 X Encoding Z However, most hierarchical and structural correlation infor- 1 1 S S ZZS 2 mation among cross-modal data cannot be well exploited due 222 to its high-dimensional non-linear property. To tackle this problem, in this paper, we propose an unsupervised frame- Z X 2 work named Cross-Modal Subspace Clustering via Deep 2 Canonical Correlation Analysis (CMSC-DCCA), which in- Different Similar corporates the correlation constraint with a self-expressive S modalities subspaces 1 ZZS layer to make full use of information among the inter-modal 111 Deep data and the intra-modal data. More specifically, the proposed Encoding X Z A better model consists of three components: 1) deep canonical corre- 1 1 S ZZS S lation analysis (Deep CCA) model; 2) self-expressive layer; Maximize 2222 3) Deep CCA decoders. The Deep CCA model consists of Correlation convolutional encoders and correlation constraint. Convolu- X Z tional encoders are used to obtain the latent representations 2 2 of cross-modal data, while adding the correlation constraint for the latent representations can make full use of the infor- Figure 1: The illustration shows the influence of correla- mation of the inter-modal data.
    [Show full text]
  • Carlos Guardia Rivas Ibm Spain
    CARLOS GUARDIA RIVAS IBM SPAIN EXECUTIVE IT SPECIALIST - IBM SOFTWARE GROUP WHAT IS THE ROLE OF DATA WITHIN YOUR ACTIVITY AT IBM? IBM is closely related to data and is a big data producer. For example, a few years ago IBM bought The Weather Company. This company generates climate data, compiles it and sells it to, especially, insurance companies and also to other private companies that make use of this kind of data. Even if IBM uses external data repositories, it is mainly a data producer. As for the tools, at IBM we have a very complete portfolio of analytics tools, ranging from importing data from diverse sources, data integration in diverse formats, data cleaning, replication to other systems if needed… Our most famous analytics tool is Watson, which is basically a set of tools that allows you to do any kind of projection on the data: from visualization to prediction. In conclusion, we are a company that offers solutions in the main data-related areas. This is the field in which IBM is heavily investing. DOES IBM USE OPEN DATA FROM PUBLIC SOURCES? We mainly use our own IBM-produced data and tools. We have a big variety of databases: structured, non-structured, analysis-oriented… CONSIDERING THAT MOST OF THE DATA IS GENERATED BY IBM, WHAT KIND OF STANDARDS DO YOU APPLY? Our own customers demand us to adjust to specific standards when they use our tools. Therefore, we comply with all the typical standards in data management. If we didn’t do so, we wouldn’t sell anything! We are always keeping an eye on the market’s demand for specific standards.
    [Show full text]
  • The Geometry of Kernel Canonical Correlation Analysis
    Max–Planck–Institut für biologische Kybernetik Max Planck Institute for Biological Cybernetics Technical Report No. 108 The Geometry Of Kernel Canonical Correlation Analysis Malte Kuss1 and Thore Graepel2 May 2003 1 Max Planck Institute for Biological Cybernetics, Dept. Schölkopf, Spemannstrasse 38, 72076 Tübingen, Germany, email: [email protected] 2 Microsoft Research Ltd, Roger Needham Building, 7 J J Thomson Avenue, Cambridge CB3 0FB, U.K, email: [email protected] This report is available in PDF–format via anonymous ftp at ftp://ftp.kyb.tuebingen.mpg.de/pub/mpi-memos/pdf/TR-108.pdf. The complete series of Technical Reports is documented at: http://www.kyb.tuebingen.mpg.de/techreports.html The Geometry Of Kernel Canonical Correlation Analysis Malte Kuss, Thore Graepel Abstract. Canonical correlation analysis (CCA) is a classical multivariate method concerned with describing linear dependencies between sets of variables. After a short exposition of the linear sample CCA problem and its analytical solution, the article proceeds with a detailed characterization of its geometry. Projection operators are used to illustrate the relations between canonical vectors and variates. The article then addresses the problem of CCA between spaces spanned by objects mapped into kernel feature spaces. An exact solution for this kernel canonical correlation (KCCA) problem is derived from a geometric point of view. It shows that the expansion coefficients of the canonical vectors in their respective feature space can be found by linear CCA in the basis induced by kernel principal component analysis. The effect of mappings into higher dimensional feature spaces is considered critically since it simplifies the CCA problem in general.
    [Show full text]
  • XLUPC: Developing an Optimizing Compiler UPC Programming Language
    IBM Software Group | Rational ® XLUPC: Developing an optimizing compiler UPC Programming Language Ettore Tiotto IBM Toronto Laboratory [email protected] ScicomP 2012 IBM Disclaimer © Copyright IBM Corporation 2012. All rights reserved. The information contained in these materials is provided for informational purposes only, and is provided AS IS without warranty of any kind, express or implied. IBM shall not be responsible for any damages arising out of the use of, or otherwise related to, these materials. Nothing contained in these materials is intended to, nor shall have the effect of, creating any warranties or representations from IBM or its suppliers or licensors, or altering the terms and conditions of the applicable license agreement governing the use of IBM software. References in these materials to IBM products, programs, or services do not imply that they will be available in all countries in which IBM operates. Product release dates and/or capabilities referenced in these materials may change at any time at IBM’s sole discretion based on market opportunities or other factors, and are not intended to be a commitment to future product or feature availability in any way. IBM, the IBM logo, Rational, the Rational logo, Telelogic, the Telelogic logo, and other IBM products and services are trademarks of the International Business Machines Corporation, in the United States, other countries or both. Other company, product, or service names may be trademarks or service marks of others. Please Note: IBM’s statements regarding its plans, directions, and intent are subject to change or withdrawal without notice at IBM’s sole discretion.
    [Show full text]
  • Lecture 7: Analysis of Factors and Canonical Correlations
    Lecture 7: Analysis of Factors and Canonical Correlations M˚ansThulin Department of Mathematics, Uppsala University [email protected] Multivariate Methods • 6/5 2011 1/33 Outline I Factor analysis I Basic idea: factor the covariance matrix I Methods for factoring I Canonical correlation analysis I Basic idea: correlations between sets I Finding the canonical correlations 2/33 Factor analysis Assume that we have a data set with many variables and that it is reasonable to believe that all these, to some extent, depend on a few underlying but unobservable factors. The purpose of factor analysis is to find dependencies on such factors and to use this to reduce the dimensionality of the data set. In particular, the covariance matrix is described by the factors. 3/33 Factor analysis: an early example C. Spearman (1904), General Intelligence, Objectively Determined and Measured, The American Journal of Psychology. Children's performance in mathematics (X1), French (X2) and English (X3) was measured. Correlation matrix: 2 1 0:67 0:64 3 R = 4 1 0:67 5 1 Assume the following model: X1 = λ1f + 1; X2 = λ2f + 2; X3 = λ3f + 3 where f is an underlying "common factor" ("general ability"), λ1, λ2, λ3 are \factor loadings" and 1, 2, 3 are random disturbance terms. 4/33 Factor analysis: an early example Model: Xi = λi f + i ; i = 1; 2; 3 with the unobservable factor f = \General ability" The variation of i consists of two parts: I a part that represents the extent to which an individual's matehematics ability, say, differs from her general ability I a "measurement error" due to the experimental setup, since examination is only an approximate measure of her ability in the subject The relative sizes of λi f and i tell us to which extent variation between individuals can be described by the factor.
    [Show full text]
  • A Brief Guide to Using This Text
    A BRIEF GUIDE TO USING THIS TEXT The sixth edition of Marketing Research offers several features that will help you throughout this course. FEATURE BENEFIT/DESCRIPTION PAGE REFERENCE SPSS and SAS computerized This edition offers the most extensive help available anywhere for learning e.g., 419, 439 demonstration movies, screen these programs. (1) From the Web site for this book you can download captures, step-by-step instruc- computerized demonstration movies illustrating step-by-step instructions for tion, and technology manual every data analysis technique covered in the chapters. (2) You can also download screen captures with notes illustrating these step-by-step instruc- tions. (3) You can study detailed step-by-step instructions in the chapters. (4) You can refer to the Study Guide and Technology Manual, a supplement that accompanies this book. We provide the most extensive help available anywhere for learning SPSS and SAS. You can learn these software tools on your own. Stimulating critical-thinking Socratic questioning, critical reading and writing, and higher-order thinking 780-787 practice and assessment are built into three comprehensive critical-thinking cases (2.1 American Idol, 2.2 Baskin-Robbins, and 2.3 Akron Children's Hospital), end-of-chapter review questions, applied problems, and group discussions. These materials have been designed along guidelines from the Foundation for Critical Thinking. Develop critical-thinking skills while you learn marketing research. Concept maps These maps will help you visualize interrelationships among concepts. These e.g., 28 are one or more concept maps per chapter based on the findings of the Institute for Human and Machine Cognition.
    [Show full text]