BRM Unit 4 Extra
BRM Unit 4 Extra
BRM Unit 4 Extra
Normality test:
The term Normality is a property of a random variable that is distributed
according to the normal distribution A normality test is used to determine whether
sample data has been drawn from a normally distributed population. Normality
tests are tests of whether a set of data is distributed in a way that is consistent with
a normal distribution.
Testing for normality is often a first step in analyzing your data. Many
statistical tools that you might use have normality as an underlying assumption. If
you fail that assumption, you may need to use a different statistical tool or
approach.
That is why, first we test the normality of the data, then we decide whether
mean is applicable as representative value of the data or not. If applicable, then
means are compared using parametric test otherwise medians are used to compare
the groups, using nonparametric methods.
Test of Uni-dimensionality:
Multicollinearity could exist because of the problems in the dataset at the time of
creation. These problems could be because of poorly designed experiments, highly
observational data, or the inability to manipulate the data:
Multicollinearity could also occur when new variables are created which are
dependent on other variables:
For example, creating a variable for BMI from the height and weight
variables would include redundant information in the model
For example, in a dataset containing the status of marriage variable with two
unique values: ‘married’, ’single’. Creating dummy variables for both of them
would include redundant information. We can make do with only one variable
containing 0/1 for ‘married’/’single’ status.
FACTOR ANALYSIS:
In exploratory factor analysis, all measured variables are related to every latent
variable. But in confirmatory factor analysis (CFA), researchers can specify the
number of factors required in the data and which measured variable is related to
which latent variable. Confirmatory factor analysis (CFA) is a tool that is used to
confirm or reject the Measurement theory.
This method is preferred by the researcher because it estimates the multiple and
interrelated dependence in a single analysis. In this analysis, two types of variables are used
endogenous variables and exogenous variables. Endogenous variables are equivalent to
dependent variables and are equal to the independent variable
Theory: This can be thought of as a set of relationships providing consistency and
comprehensive explanations of the actual phenomena. There are two types of models:
Measurement model: The measurement model represents the theory that specifies how measured
variables come together to represent the theory.
Structural model: Represents the theory that shows how constructs are related to other
constructs. Structural equation modelling is also called casual modelling because it tests the
proposed casual relationships. The following assumptions are assumed:
Multivariate normal distribution: The maximum likelihood method is used and assumed for
multivariate normal distribution. Small changes in multivariate normality can lead to a large
difference in the chi-square test.
Outlier: Data should be free of outliers. Outliers affect the model significance.
Sequence: There should be a cause and effect relationship between endogenous and exogenous
variables, and a cause has to occur before the event.
Model identification: Equations must be greater than the estimated parameters or models should
be over identified or exact identified. Under identified models are not considered.
Sample size: Most of the researchers prefer a 200 to 400 sample size with 10 to 15 indicators.
As a rule of thumb, that is 10 to 20 times as many cases as variables.
Uncorrelated error terms: Error terms are assumed uncorrelated with other variable error
terms.
CONJOINT ANALYSIS
Conjoint analysis is a form of statistical analysis that firms use in market research
to understand how customers value different components or features of their products or
services. It’s based on the principle that any product can be broken down into a set of
attributes that ultimately impact users’ perceived value of an item or service.
The insights a company gleans from conjoint analysis of its product features can
be leveraged in several ways. Most often, conjoint analysis impacts pricing strategy, sales
and marketing efforts, and research and development plans
Choice-Based Conjoint (CBC) Analysis: This is one of the most common forms of
conjoint analysis and is used to identify how a respondent values combinations of
features.
Adaptive Conjoint Analysis (ACA): This form of analysis customizes each
respondent's survey experience based on their answers to early questions. It’s often
leveraged in studies where several features or attributes are being evaluated to
streamline the process and extract the most valuable insights from each respondent.
Full-Profile Conjoint Analysis: This form of analysis presents the respondent with a
series of full product descriptions and asks them to select the one they’d be most
inclined to buy.
MaxDiff Conjoint Analysis: This form of analysis presents multiple options to the
respondent, which they’re asked to organize on a scale of “best” to “worst” (or “most
likely to buy” to “least likely to buy”)
Application of Statistical Software for Data Analysis:
The key benefit to using a spreadsheet program is that you can make changes easily,
including correcting spelling or values, adding, deleting, formatting, and relocating data. You
can also program the spreadsheet to perform certain functions automatically (such as addition
and subtraction), and a spreadsheet can hold almost limitless amounts of data—a whole filing
cabinet’s worth of information can be included in a single spreadsheet.
2) SPSS
SPSS means “Statistical Package for the Social Sciences” and was first launched in
1968. Since SPSS was acquired by IBM in 2009, it's officially known as IBM SPSS
Statistics but most users still just refer to it as “SPSS
SPSS is software for editing and analyzing all sorts of data. These data may come from
basically any source: scientific research, a customer database, Google Analytics or even the
server log files of a website.
3) AMOS
AMOS is statistical software and it stands for analysis of a moment structures. AMOS
is an added SPSS module, and is specially used for Structural Equation Modeling, path
analysis, and confirmatory factor analysis.
It is also known as analysis of covariance or causal modeling software. AMOS is a
visual program for structural equation modeling (SEM). In AMOS, we can draw models
graphically using simple drawing tools. AMOS quickly performs the computations for
SEM and displays the results
4) R-analysis
R analytics is data analytics using R programming language, an open-source language used for
statistical computing or graphics. This programming language is often used in statistical analysis
and data mining.
It can be used for analytics to identify patterns and build practical models. R not only can help
analyze organizations’ data, but also be used to help in the creation and development of software
applications that perform statistical