Get AI summaries of any video or article — Sign up free
LESSON 67 - RESEARCH METHODOLOGY || SECTION 3.8: DATA ANALYSIS TECHNIQUES || QUANTITATIVE DATA thumbnail

LESSON 67 - RESEARCH METHODOLOGY || SECTION 3.8: DATA ANALYSIS TECHNIQUES || QUANTITATIVE DATA

5 min read

Based on RESEARCH METHODS CLASS WITH PROF. LYDIAH WAMBUGU's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Section 3.8 should specify how collected quantitative (numerical) data will be analyzed, including the statistical tools and the order of presentation.

Briefing

Quantitative data analysis in a research proposal should be presented as a clear, step-by-step plan: reduce raw responses into manageable summaries, then use statistics to test hypotheses and generalize from a sample to a population. The core idea is that once questionnaires or other numerical instruments generate data, the researcher must specify how that data will be analyzed, what statistical tools will be used, and how results will be organized and presented—typically starting with descriptive statistics and moving to inferential statistics.

Data analysis is defined as the process of reducing collected data into meaningful summaries. In most social science dissertations, this work is largely carried out in Chapter 4, where the collected responses are summarized according to variables. The choice of analysis method depends on the type of data and the scale of measurement. Numerical data calls for statistical analysis, while narrative data requires different approaches (handled in later lessons). For quantitative work, manual analysis is no longer the norm; common tools include IBM SPSS for quantitative/numerical data and NVivo for narrative/qualitative data.

The lesson then separates quantitative analysis into two major categories. Descriptive statistics summarize a sample by using one number to represent a group of numbers. Inferential statistics go further: they support predictions and generalizations about the population based on sample data, relying on probability and hypothesis testing. Hypothesis testing splits into parametric tests—built on assumptions such as the population being normally distributed—and non-parametric tests, often described as “assumption-free” because they make fewer assumptions and frequently use ranking (analyzing ranks rather than raw values).

For presenting descriptive results, three organizing approaches are emphasized: graphical, tabular, and numerical summaries. Graphical methods include histograms, cutter diagrams, and frequency polygons, with specific guidance that some visuals suit categorical data while others suit continuous data. Tabular presentation includes frequency distribution tables and crosstabs, particularly for categorical variables. Numerical summaries include measures of central tendency and variability—mean, median, variance, and standard deviation—typically for continuous data.

Inferential statistics are also grouped into three families. Correlation examines relationships: Spearman rank-order correlation for ordinal data and chi-square for nominal categorical data, while Pearson product-moment correlation coefficient is used for continuous data. Regression is used for continuous outcomes, including simple and multiple regression, with binary logistic regression mentioned for earlier lessons tied to the type of data. Tests of comparison include t-tests (noted as having three types) and analysis of variance (ANOVA), both applied to continuous data.

Finally, the lesson lays out what researchers should write in Section 3.8: first identify the statistical tools for quantitative data, then explain how those tools answer the research questions and test hypotheses, and then describe the order of presentation—descriptives first (tabular/graphical/numerical), followed by inferential results. The next session shifts from quantitative to qualitative data analysis techniques.

Cornell Notes

Quantitative data analysis is the plan for turning collected numerical responses into summaries and statistical conclusions. The workflow typically starts with descriptive statistics (summarizing the sample) and then moves to inferential statistics (generalizing to the population using probability and hypothesis testing). Hypothesis testing uses parametric tests when normality assumptions are reasonable, and non-parametric tests when fewer assumptions are preferred, often relying on ranking. Descriptive results can be presented graphically, in tables (frequency distributions and crosstabs), or numerically (mean, median, variance, standard deviation). Inferential tools are grouped into correlation (Spearman, chi-square, Pearson), regression (simple/multiple, plus binary logistic for earlier binary cases), and comparison tests (t-tests and ANOVA).

Why does Section 3.8 require both descriptive and inferential statistics in quantitative research proposals?

Descriptive statistics summarize the sample by using numbers to represent patterns in the collected data. Inferential statistics support predictions and generalizations about the population based on sample results, and they do this through probability-based hypothesis testing. That link—sample summary first, then population inference—drives the structure of Section 3.8.

How should a researcher decide between parametric and non-parametric tests?

Parametric tests assume the population is normally distributed (assumptions about population parameters). Non-parametric tests make fewer assumptions and are often described as “assumption-free,” frequently using ranking principles—analyzing ranks rather than the raw data values.

What are the main ways to present descriptive statistics for quantitative data?

Descriptive results can be presented graphically (e.g., histograms, cutter diagrams, frequency polygons), tabularly (frequency distribution tables and crosstabs), or numerically (mean, median, variance, standard deviation). The choice aligns with the data type, with categorical variables commonly handled through frequency/crosstab tables and continuous variables through measures like mean/standard deviation and continuous-data graphs.

Which inferential methods are grouped under correlation, and what data types do they match?

Correlation is split by variable type: Spearman rank-order correlation is for ordinal data, chi-square is for nominal categorical data, and Pearson product-moment correlation coefficient is for continuous data. Each method matches the measurement level and the kind of relationship being tested.

How do regression and comparison tests fit into quantitative analysis?

Regression is used for continuous outcomes, including simple and multiple regression, and binary logistic regression is mentioned for binary cases handled in earlier lessons. Comparison tests include t-tests (with three types) and analysis of variance (ANOVA), both applied to continuous data when comparing groups or conditions.

What is the recommended order for writing and organizing quantitative analysis in a proposal?

After identifying the statistical tools and explaining how they answer the research questions and test hypotheses, the presentation should follow a clear sequence: start with descriptive statistics (tabular/graphical/numerical), then move to inferential statistics. This order helps readers see how raw data becomes summaries, then conclusions.

Review Questions

  1. In your own words, distinguish descriptive statistics from inferential statistics and give one reason each is needed in Section 3.8.
  2. Match each method to a likely data type: Spearman rank-order, chi-square, Pearson product-moment, and ANOVA.
  3. Outline a Section 3.8 paragraph structure that starts with tool identification, then hypothesis testing, then the order of data presentation.

Key Points

  1. 1

    Section 3.8 should specify how collected quantitative (numerical) data will be analyzed, including the statistical tools and the order of presentation.

  2. 2

    Data analysis is the process of reducing raw responses into manageable summaries that can be presented, interpreted, and used to answer research questions.

  3. 3

    Descriptive statistics summarize the sample, while inferential statistics generalize to the population using probability and hypothesis testing.

  4. 4

    Hypothesis tests split into parametric tests (normality assumptions) and non-parametric tests (fewer assumptions, often rank-based).

  5. 5

    Descriptive results should be organized as graphical, tabular, and numerical outputs, with methods aligned to categorical versus continuous data.

  6. 6

    Inferential statistics can be grouped into correlation, regression, and tests of comparison, each matched to the measurement level and outcome type.

  7. 7

    A strong quantitative analysis write-up starts with descriptives and then proceeds to inferential results after tool selection and hypothesis-testing explanation.

Highlights

Quantitative analysis should be written as a plan: identify statistical tools, explain how they test hypotheses, then present results in a logical sequence—descriptives first, inferentials second.
Inferential statistics enable generalization to the population because they rely on probability and hypothesis testing, not just sample summaries.
Parametric tests assume normally distributed populations, while non-parametric tests reduce assumptions and often work with ranks rather than raw values.
Descriptive statistics can be presented graphically (histograms, frequency polygons), in tables (frequency distributions, crosstabs), or numerically (mean, median, variance, standard deviation).
Correlation, regression, and comparison tests form the core families of inferential statistics, each tied to specific data types (ordinal, nominal, continuous).

Topics