You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Given a real-time operational constraints, this library can
automatically compress the network to satisfy the constraint while
preserving the accuracy. This library is built on top of SkimCaffe,
which has implemented direct sparse convolution operations and has an effective
speedup given an input sparse network. This framework can be applied to
different constraint types(latency, memory size), different network(Alexnet,
Resnet, Googlenet) and different datasets(ImagenNet, DTD).
For more technical details, please refer to the paper.
Usage
First follow the build-up instructions in SkimCaffe
to build SkimCaffe.
Modify the cfp.config file to adapt to your own need(constraint values).
Run training in root directory:
python2 main.py
To visualize the training result, specify the output directory:
Visualization of sampled data points in the first exponential cooling step
Citation
If you use this code or ideas from the paper for your research, please cite our paper:
@InProceedings{Chen_2018_ECCV,
author = {Chen, Changan and Tung, Frederick and Vedula, Naveen and Mori, Greg},
title = {Constraint-Aware Deep Neural Network Compression},
booktitle = {The European Conference on Computer Vision (ECCV)},
month = {September},
year = {2018}
}
About
[ECCV18] Constraint-Aware Deep Neural Network Compression