Skip to content

Rising-Stars-by-Sunshine/Yuxuan-Microexpression

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

43 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Predicting Emotion Intensity and Addressing Class Imbalance in Micro-Expression Recognition

Abstract

This project investigates two predictive challenges in micro-expression recognition using open-source Kaggle datasets: (1) extending traditional categorical classification to continuous emotion intensity prediction, and (2) improving recognition of minority emotion classes (e.g., fear, disgust) under imbalanced data conditions.
By comparing regression- and classification-based approaches, and by applying imbalance-handling methods such as SMOTE, class weights, and Focal Loss, the project aims to improve both the granularity and the fairness of emotion recognition models.
The study integrates the AI Triad (data, algorithms, computing power) and reflects critically on the role of GenAI tools (ChatGPT, STORM, Hugging Face) in accelerating reproducible, ethically responsible research.


System Configuration

  • Local setup:
    • Python 3.10+
    • Jupyter Notebook
    • PyTorch / TensorFlow (for deep learning experiments)
    • scikit-learn, imbalanced-learn (for baseline models and imbalance handling)
  • Cloud setup:
    • Google Colab (GPU runtime for model training)
    • Google Drive (dataset hosting and integration)
    • Hugging Face Hub (pretrained vision models for transfer learning)

Research Framing & AI Triad Connections

flowchart A[Research Question] --> B1[Emotion Intensity Prediction] A --> B2[Minority Class Prediction]

B1 --> C1[Data: Kaggle datasets - soft labels 0-1]
B1 --> D1[Algorithms: CNN + LSTM; Regression loss (MSE/MAE)]
B1 --> E1[Compute: Colab GPU for training]

B2 --> C2[Data: Imbalanced micro-expression samples]
B2 --> D2[Algorithms: CNN + LSTM with imbalance handling (SMOTE, Focal Loss, Class weights)]
B2 --> E2[Compute: Colab GPU + scikit-learn baselines]

A --> F[Integration of GenAI Tools]
F --> G1[ChatGPT: code prototyping]
F --> G2[STORM: literature mapping]
F --> G3[Hugging Face: pretrained models]

FAIR & CARE Principles

FAIR: All datasets used are open-access and properly cited. Code, preprocessing scripts, and notebooks are shared in this repository with clear documentation. Outputs are stored in interoperable formats (CSV, JSON, PNG).

CARE: Collective benefit: Research aims to support education and healthcare applications. Authority to control: Dataset licensing and participant consent are acknowledged. Responsibility: Results emphasize transparency and report limitations, avoiding misuse in surveillance contexts. Ethics: Bias and fairness are critically evaluated, particularly regarding minority emotion classes.

📘 Notebooks

  1. Prediction_of__Microexpression_basic_EDA_ipynb_.ipynb Performs basic exploratory data analysis (EDA) on datasets. Includes class distribution plots and imbalance checks. Provides baseline evaluation metrics (Accuracy, Macro-F1, Weighted-F1).

  2. Explanation_NLP_network_analysis.ipynb Uses NLP (word cloud, keyword frequency) and network analysis (semantic co-occurrence networks) to map research themes. Identifies central and peripheral terms in micro-expression literature. Connects findings to broader research challenges (dataset bias, subtlety of low-frequency emotions).

Results Summary

Baseline (ResNet18, no SMOTE): High accuracy but biased toward majority classes. PCA + SMOTE: Lower accuracy, but better recall and Macro-F1 for minority classes (fear, disgust). AutoML (FLAML): Best overall accuracy but less interpretable. Logistic Regression: Competitive performance with highest interpretability.

Acknowledgments and Revisions

This project benefited greatly from the constructive feedback of reviewers, whose comments guided substantial improvements in both methodology and presentation. Specifically:

Reviewer Feedback Integration Expanded rationale for selecting ResNet18 compared with other CNN architectures (ResNet50, VGG, MobileNet), clarifying the trade-off between computational efficiency and feature richness. Enhanced dataset transparency by explicitly listing the three datasets used, documenting Google Images sourcing protocols, and adding a summary table of dataset characteristics. Improved terminology clarity with explicit definitions and citations for Macro-F1, Weighted-F1, PCA, SMOTE, and CNN features. Strengthened ethical safeguards by specifying informed consent procedures, withdrawal rights, cultural adaptation of materials, and requirements for bias auditing in deployed systems. Added a System Configuration section in the README for full reproducibility (Python version, Colab environment, dependencies). Expanded figure captions to highlight logical relationships between technical methods, challenges, and ethical concerns.

Acknowledgments I would like to thank Prof. Zhang and Jingting for their thoughtful and detailed reviews, which significantly improved the rigor, clarity, and ethical grounding of this work. Their feedback shaped both the updated manuscript and this repository’s documentation.

These revisions ensure that the project is methodologically transparent, ethically responsible, and aligned with both FAIR and CARE principles.

About

Yuxuan Huang-PS1

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 100.0%