Research Methods for User Studies - DSCI-517

Posted by Vineet Kumar Loyer on Thursday, February 20, 2025

1. Need for Data Science


Volume, variety, and velocity of data are growing at an unprecedented rate. Data science has emerged as a discipline that brings together methods from statistics, machine learning, and domain expertise to extract actionable insights from this diverse data.

Key points touched in lecture -

  • Data Explosion: With the advent of social media, sensors, digital transactions, and mobile devices, enormous amounts of data are generated every minute. This data is heterogeneous, encompassing text, images, geospatial coordinates, time series, and more.
  • Complex Problems: Traditional analysis methods may be insufficient to tackle the complexity of modern data. Data science provides new frameworks and computational tools to address issues like predicting market trends, improving public safety, or optimizing logistical operations.
  • Decision-Making: In both industry and academia, data-driven decisions lead to better outcomes. Whether predicting stock market fluctuations based on public sentiment or designing emergency response systems using geospatial data, data science is central to making informed, reliable decisions.

2. Framework for Data Science


Two central methodologies -

Deduction (“Top-Down” Approach):

  • Definition: Deduction starts with a general theory or hypothesis and moves toward specific predictions. In other words, one begins with established principles and derives specific consequences that can be tested with data.
  • Example in Practice: Consider a scenario where a researcher uses established economic theory to hypothesize that a decrease in interest rates will lead to increased consumer spending. Statistical tests are then applied to relevant financial data to verify this prediction.
  • Statistical Connection: This approach is strongly related to hypothesis testing, where a null hypothesis is set up and then either rejected or not rejected based on the observed data (p-values, significance levels, etc.).

Induction (“Bottom-Up” Approach):

  • Definition: Induction moves from specific observations to broader generalizations. It is about gathering data and then identifying patterns or rules that emerge from that data.
  • Example in Practice: Using customer transaction data to identify common buying patterns without an a priori hypothesis about consumer behavior is a classic inductive approach. For instance, analyzing purchase history to determine that customers who buy product A often also buy product B.
  • Statistical Connection: This approach is inherent in exploratory data analysis, where statistical methods (like clustering or regression analysis) are used to find patterns that can later be generalized into a theory.

Note: Both are critical - Deductive methods provide a clear pathway to testing established theories, while inductive methods allow for the discovery of new insights from complex datasets.

3. Types of Data


3.1 Text Data

  • Overview: Text data is unstructured and includes sources like tweets, blogs, emails, and other written communications.
  • Key Characteristics:
    • Natural Language: Requires techniques from natural language processing (NLP) to convert text into quantifiable data (tokenization, sentiment analysis, topic modeling).
    • Applications:
      • Public Sentiment Analysis: For instance, analyzing Twitter data to gauge public mood can help predict economic trends such as stock market movements. Studies have shown that shifts in public sentiment captured through text can serve as early indicators of market volatility.
      • Political Analysis: Analyzing comments and debates on social media or blogs can forecast election outcomes or measure public opinion on policies.
  • Example Explanation: Imagine a scenario where researchers use a large corpus of tweets to automatically measure the public mood about the economy. By applying sentiment analysis algorithms, they can detect shifts in tone (e.g., more negative sentiment during economic downturns) and correlate these changes with stock market indices like NASDAQ or the S&P 500. This approach leverages both statistical methods (to validate correlations) and machine learning (to classify sentiment) to provide actionable insights.

3.2 Geospatial Data

  • Overview: Geospatial data involves any information tied to specific geographic locations.
  • Key Characteristics:
    • Multiple Levels of Precision: Data can range from precise GPS coordinates (exact addresses) to broader categorizations (city, state, or country).
    • Integration with Other Data Types: Often combined with text data (e.g., location-based tweets) for richer analysis.
  • Applications:
    • Emergency Response: Geotagged social media posts can help in identifying and responding to disasters like fires or earthquakes. For example, a sudden spike in tweets containing keywords such as “fire” or “earthquake” in a specific region can trigger emergency services.
    • Urban Planning: Analyzing movement patterns in cities to optimize transportation or public service delivery.
  • Example Explanation: Consider a scenario where emergency responders use a combination of geospatial data and social media analysis. When a disaster occurs, responders can quickly pinpoint the affected area by analyzing geotagged tweets. The integration of different levels of location precision (exact coordinates from mobile devices versus city-level data from user profiles) improves the accuracy of event detection and response coordination.

3.3 Multimedia Data: Images and Videos

  • Images:
    • Overview: Multimedia data includes images that are rich in visual information. Unlike text, image data consists of pixels, colors, textures, and shapes.
    • Applications:
      • Object Recognition: Identifying objects within an image (e.g., distinguishing between different types of buildings or recognizing handwritten digits).
      • Pre-processing Techniques: Methods like Otsu’s method for image thresholding are used to differentiate between the object and the background.
    • Example Explanation: In an image recognition task, an algorithm might be trained to differentiate between pictures of windows, roofs, or houses. The computer “sees” these images as arrays of pixels, and using edge-detection algorithms, it can identify the shapes and patterns that correspond to specific objects.
  • Videos:
    • Overview: Video data adds the dimension of time to images, creating a sequence of frames that can be analyzed for motion and activity.
    • Applications:
      • Object Tracking: Beyond recognizing objects, video analysis involves tracking their movement over time. For example, a surveillance system may track a car as it moves through a frame.
      • Activity Recognition: This combines object recognition and tracking to identify activities, such as a person running or a vehicle turning.
    • Example Explanation: In a traffic surveillance system, the video analysis might first detect vehicles (object recognition), then follow them as they move along a road (object tracking), and finally identify patterns such as congestion or accidents (activity recognition). Such systems are crucial for urban traffic management and public safety.

3.4 Network Data

  • Overview: Network data represents relationships or connections between entities. These could be social networks, transportation networks, or even biological networks.
  • Key Characteristics:
    • Nodes and Edges: Entities are represented as nodes and their relationships as edges.
    • Scale-Free Networks: A common characteristic of many networks is that a few nodes (hubs) have a vast number of connections, while most nodes have only a few.
  • Applications:
    • Social Network Analysis: Visualizing and analyzing networks such as the co-sponsorship patterns among U.S. senators can reveal underlying political alliances or ideological clusters.
    • Infrastructure Networks: Studying networks like the Internet, railroad connections, or flight routes to understand connectivity and resilience.
  • Example Explanation: An example from the slides is TouchGraph’s visualization of U.S. senators’ co-sponsorship patterns. By mapping the connections, one can observe that most senators have relatively few connections while a select few (hubs) are highly connected. This scale-free property is crucial for understanding the dynamics of information flow and influence within the network.

3.5 Tabular Data

  • Overview: Tabular data is highly structured and consists of rows and columns—think of spreadsheets or relational databases.
  • Key Characteristics:
    • Cell Accessibility: Unlike images or PDFs, tabular data allows each cell to be individually accessed and analyzed.
    • Structured Format: This type of data is ideal for statistical analysis, where each column represents a variable (e.g., age, income) and each row represents an observation.
  • Applications:
    • Statistical Analysis: Regression, hypothesis testing, and descriptive statistics are often performed on tabular data.
    • Predictive Modeling: In many machine learning tasks, tabular data serves as the input for classification or regression models.
  • Example Explanation: When working with tabular data, it is critical that the data be “clean” and well-structured so that each variable can be analyzed independently. For example, in a clinical trial dataset, one might have columns for treatment type, patient age, and recovery rate. Statistical tests can then determine if the treatment has a significant effect on recovery.

3.6 Time Series Data

  • Overview: Time series data is collected sequentially over time, making the time component an integral aspect of the data.
  • Key Characteristics:
    • Temporal Order: Each observation is time-stamped, which is crucial for identifying trends, seasonal effects, or cycles.
    • Analysis Techniques: Methods like moving averages, exponential smoothing, or autoregressive models are commonly used.
  • Applications:
    • Weather Forecasting: Data such as temperature or precipitation recorded over time helps in building predictive weather models.
    • Economic Indicators: Time series data is used to track stock prices, unemployment rates, or other financial metrics.
  • Example Explanation: weather station data, where temperatures are recorded at hourly intervals. Analysts might use such data to identify diurnal temperature variations or to forecast future weather conditions by modeling the observed time series data using statistical methods.

4. Types of Analysis


Once data is collected, the next step in data science is analysis. The slide deck outlines several methods of analysis, each with distinct approaches and use cases. Here, we explore these techniques and provide examples to illustrate their applications.

4.1 Machine Learning and Other Analyses

Data analysis in data science can broadly be divided into machine learning (ML) and other statistical methods such as simulation and hypothesis testing. Machine learning itself can be further divided into:

  • Supervised Learning: The model is trained on labeled data, meaning that the input data is accompanied by the correct output (or label). The goal is to learn a mapping from inputs to outputs.
  • Unsupervised Learning: The model is given unlabeled data and must find structure or patterns in the data on its own.

4.2 Classification

  • Definition: Classification is a supervised learning task where the goal is to assign a new instance to one of several predefined categories or classes.
  • Statistical Basis: Classification relies on statistical algorithms to build predictive models from labeled examples. Methods such as logistic regression, decision trees, or support vector machines are common.
  • Example – Classifying Mushrooms:
    • Problem Statement: Determine whether a given mushroom is edible or poisonous based on its features.
    • Features: The slide lists several features used in the mushroom dataset, such as cap shape, cap surface, cap color, bruises, odor, gill characteristics, stalk shape, stalk root, and so on.
    • Explanation: Each mushroom is represented by a vector of feature values (e.g., cap shape might be “bell” or “conical”, odor might be “almond” or “foul”). A classification algorithm is trained on a set of mushrooms whose edibility is known (edible vs. poisonous). Once the model is built, it can predict the class of a new, unseen mushroom by comparing its feature vector to the learned patterns.
    • Statistical Considerations: Feature selection, handling categorical variables (often through one-hot encoding), and cross-validation to avoid overfitting are all critical steps in building a robust classification model.

4.3 Clustering

  • Definition: Clustering is an unsupervised learning method that groups a set of instances so that those in the same group (cluster) are more similar to each other than to those in other clusters.
  • Techniques: Algorithms such as k-means clustering, DBSCAN, or hierarchical clustering are commonly used.
  • Example Explanation: Suppose we have a dataset of customer transactions without predefined labels. Clustering can be used to segment customers into distinct groups based on purchasing behavior. These segments can then inform targeted marketing strategies or customer service approaches. The optimization criterion here is typically to minimize within-cluster variance, ensuring that each cluster is as homogeneous as possible.

4.4 Pattern Analysis

  • Overview: Pattern analysis involves identifying regularities or repeated structures within the data. This can be divided into:
    • Pattern Detection/Learning: Given data that is already annotated with known patterns, the goal is to detect matches.
    • Pattern Discovery/Mining: Finding new, previously unknown patterns in the data.
  • Example Explanation: In retail, pattern discovery might reveal that sales of umbrellas and raincoats spike concurrently during specific weather conditions. Statistical association rules (such as those used in market basket analysis) can be applied to uncover these relationships.

4.5 Simulation

  • Definition: Simulation involves constructing mathematical models that mimic real-world processes. These models can generate synthetic data under different scenarios, which can then be compared to observed data.
  • Applications:
    • Traffic Simulation: By modeling the flow of vehicles, simulations can predict traffic congestion, evaluate the impact of infrastructure changes, or optimize traffic signal timings.
    • Engineering and Natural Sciences: Simulations of airflow over an engine or climate models in hydrology are prime examples.
  • Statistical Underpinnings: Simulation models often rely on statistical distributions and Monte Carlo methods to generate synthetic data that reflects the variability observed in real-world processes.

4.6 Null Hypothesis Significance Testing

  • Overview: Null hypothesis significance testing (NHST) is a statistical framework for testing whether an observed effect in a sample can be attributed to chance.
  • Key Components:
    • Null Hypothesis (H₀): The assumption that there is no effect or relationship between variables.
    • Alternative Hypothesis (H₁): The hypothesis that there is an effect or relationship.
    • p-value: The probability of obtaining the observed result, or something more extreme, if the null hypothesis is true.
  • Example Explanation: Imagine a clinical trial where the effectiveness of a new drug is being evaluated. The null hypothesis might state that the drug has no effect on patient recovery rates. Statistical tests (e.g., t-tests or ANOVA) are then performed, and if the p-value is below a threshold (commonly 0.05), the null hypothesis is rejected, suggesting that the drug has a statistically significant effect.
  • Statistical Relevance: NHST is fundamental in many scientific disciplines as it provides a formal mechanism to decide whether observed data supports a particular theoretical mode