Skip to content

DensoITLab/sas_

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

32 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

SAS: Structured Activation Sparsification

This is the official repo for ICLR 2024 Paper "SAS: Structured Activation Sparsification" Yusuke Sekikawa and Shingo Yashima

paper, openreview

Overview

sas_alg

SAS expands representational capacity by projecting inputs into a higher-dimensional, structured-sparse space.This is equivalent to selecting the appropriate weights depending on its input—thereby improving accuracy without increasing FLOPs.

line_chart

All models were trained on the ImageNet dataset under the following common settings:

  • Loss function: Distillation loss
  • Data augmentation: AutoAugment
  • Batch size: 256
  • Learning rate scheduler: Cosine annealing
  • Optimizer: SGD (momentum=0.9, weight decay=1e-4)

※ Wide ResNet-18 refers to a model that has twice as many parameters as ResNet-18.

Computation

flops

  • SAS achieves a 46.0% reduction in FLOPs at a fixed Top-1 accuracy.
    • The width (X≒1.8) that matches SAS’s Top-1 accuracy was determined by linearly interpolating between the Top-1 accuracies of ResNet-18 and Wide ResNet-18. We then set this Wide (X≒1.8) ResNet-18 as the 100 % FLOPs baseline and calculated the relative FLOPs of SAS accordingly.

Accuracy

acc

  • SAS increases Top-1 accuracy by 2.63% at the same FLOPs.

Run

1. Train standard ResNet-18

python train.py <TRAIN_DATA_DIR> <VAL_DATA_DIR>

2. Train ResNet-18 with SAS

python train.py <TRAIN_DATA_DIR> <VAL_DATA_DIR> --use_sas

3. Train Wide ResNet-18

python train.py <TRAIN_DATA_DIR> <VAL_DATA_DIR> --arch wide_resnet18

4. Train Wide ResNet-18 with SAS

python train.py <TRAIN_DATA_DIR> <VAL_DATA_DIR> --use_sas --arch wide_resnet18

Citation

If you find our code or paper useful, please cite the following:

@inproceedings{
sekikawa2024sas,
title={{SAS}: Structured Activation Sparsification},
author={Yusuke Sekikawa and Shingo Yashima},
booktitle={The Twelfth International Conference on Learning Representations},
year={2024},
url={https://openreview.net/forum?id=vZfi5to2Xl}
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •  

Languages