← Back to Projects

Saddle-to-Saddle Dynamics with Simplicity Bias

saddle-to-saddle_dynamics_with_simplicity_bias_can Not Started

Project Actions

Open in Terminal

Project Status

Source Idea

Saddle-to-saddle dynamics with simplicity bias can be leveraged to improve the training efficiency of neural networks by dynamically adjusting learning rates during training.

View Source Idea →

Files (7)

  • README.md
  • metadata.json
  • requirements.txt
  • src/evaluate.py
  • src/models.py
  • src/train.py
  • src/utils.py

README Preview

# Saddle-to-Saddle Dynamics with Simplicity Bias ## Description This project investigates the hypothesis that saddle-to-saddle dynamics with simplicity bias can enhance the training efficiency of neural networks by dynamically adjusting learning rates during training. ## Research Hypothesis Saddle-to-saddle dynamics with simplicity bias can be leveraged to improve the training efficiency of neural networks by dynamically adjusting learning rates during training. ## Implementation Approach The implementation involves developing a dynamic learning rate scheduler that integrates saddle-to-saddle dynamics. Experiments will be conducted on various datasets and neural network architectures to evaluate the impact on training efficiency and generalization. ## Setup Instructions 1. Clone the repository. 2. Install the required dependencies using `pip install -r requirements.txt`. 3. Prepare the datasets (CIFAR-10, MNIST) in the data directory. ## Usage Run the training script: ```bash python src/train.py --dataset CIFAR-10 --model resnet18 ``` Evaluate the model: ```bash python src/evaluate.py --model checkpoints/model.pth ``` ## Expected Results The project aims to demonstrate improved training efficiency and generalization performance compared to traditional learning rate schedules. ## References - Saddle-to-Saddle Dynamics Explains A Simplicity Bias Across Neural Network Architectures [arXiv:2512.20607v1](http://arxiv.org/abs/2512.20607v1)