Densely Connected Convolutional Networks
- Gao Huang, Zhuang Liu, Kilian Q. Weinberger
- Computer ScienceComputer Vision and Pattern Recognition
- 25 August 2016
The Dense Convolutional Network (DenseNet), which connects each layer to every other layer in a feed-forward fashion, and has several compelling advantages: they alleviate the vanishing-gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters.
Learning Efficient Convolutional Networks through Network Slimming
- Zhuang Liu, Jianguo Li, Zhiqiang Shen, Gao Huang, Shoumeng Yan, Changshui Zhang
- Computer ScienceIEEE International Conference on Computer Vision
- 22 August 2017
The approach is called network slimming, which takes wide and large networks as input models, but during training insignificant channels are automatically identified and pruned afterwards, yielding thin and compact models with comparable accuracy.
Deep Networks with Stochastic Depth
- Gao Huang, Yu Sun, Zhuang Liu, Daniel Sedra, Kilian Q. Weinberger
- Computer ScienceEuropean Conference on Computer Vision
- 30 March 2016
Stochastic depth is proposed, a training procedure that enables the seemingly contradictory setup to train short networks and use deep networks at test time and reduces training time substantially and improves the test error significantly on almost all data sets that were used for evaluation.
A ConvNet for the 2020s
- Zhuang Liu, Hanzi Mao, Chaozheng Wu, Christoph Feichtenhofer, Trevor Darrell, Saining Xie
- Computer ScienceComputer Vision and Pattern Recognition
- 10 January 2022
This work gradually “modernize” a standard ResNet toward the design of a vision Transformer, and discovers several key components that contribute to the performance difference along the way, leading to a family of pure ConvNet models dubbed ConvNeXt.
Rethinking the Value of Network Pruning
- Zhuang Liu, Mingjie Sun, Tinghui Zhou, Gao Huang, Trevor Darrell
- Computer ScienceInternational Conference on Learning…
- 27 September 2018
It is found that with optimal learning rate, the "winning ticket" initialization as used in Frankle & Carbin (2019) does not bring improvement over random initialization, and the need for more careful baseline evaluations in future research on structured pruning methods is suggested.
Few-Shot Object Detection via Feature Reweighting
- Bingyi Kang, Zhuang Liu, Xin Wang, F. Yu, Jiashi Feng, Trevor Darrell
- Computer ScienceIEEE International Conference on Computer Vision
- 5 December 2018
This work develops a few-shot object detector that can learn to detect novel objects from only a few annotated examples, using a meta feature learner and a reweighting module within a one-stage detection architecture.
Snapshot Ensembles: Train 1, get M for free
- Gao Huang, Yixuan Li, Geoff Pleiss, Zhuang Liu, J. Hopcroft, Kilian Q. Weinberger
- Computer ScienceInternational Conference on Learning…
- 1 April 2017
This paper proposes a method to obtain the seemingly contradictory goal of ensembling multiple neural networks at no additional training cost by training a single neural network, converging to several local minima along its optimization path and saving the model parameters.
DSOD: Learning Deeply Supervised Object Detectors from Scratch
- Zhiqiang Shen, Zhuang Liu, Jianguo Li, Yu-Gang Jiang, Yurong Chen, X. Xue
- Computer ScienceIEEE International Conference on Computer Vision
- 3 August 2017
Deeply Supervised Object Detector (DSOD), a framework that can learn object detectors from scratch following the single-shot detection (SSD) framework, and one of the key findings is that deep supervision, enabled by dense layer-wise connections, plays a critical role in learning a good detector.
Test-Time Training with Self-Supervision for Generalization under Distribution Shifts
- Yu Sun, X. Wang, Zhuang Liu, John Miller, Alexei A. Efros, Moritz Hardt
- Computer ScienceInternational Conference on Machine Learning
- 29 September 2019
This work turns a single unlabeled test sample into a self-supervised learning problem, on which the model parameters are updated before making a prediction, which leads to improvements on diverse image classification benchmarks aimed at evaluating robustness to distribution shifts.
A New Meta-Baseline for Few-Shot Learning
- Yinbo Chen, Xiaolong Wang, Zhuang Liu, Huijuan Xu, Trevor Darrell
- Computer ScienceArXiv
- 9 March 2020
This work presents a Meta-Baseline method, by pre-training a classifier on all base classes and meta-learning on a nearest-centroid based few-shot classification algorithm, which outperforms recent state-of-the-art methods by a large margin.
...
...