0 |
ZF Net (Visualizing and Understanding Convolutional Networks) |
Read |
CNNs, CV , Image |
Visualization |
2014 |
ECCV |
Matthew D. Zeiler, Rob Fergus |
Visualize CNN Filters / Kernels using De-Convolutions on CNN filter activations. |
link |
1 |
Inception-v1 (Going Deeper With Convolutions) |
Read |
CNNs, CV , Image |
Architecture |
2015 |
CVPR |
Christian Szegedy, Wei Liu |
Propose the use of 1x1 conv operations to reduce the number of parameters in a deep and wide CNN |
link |
2 |
ResNet (Deep Residual Learning for Image Recognition) |
Read |
CNNs, CV , Image |
Architecture |
2016 |
CVPR |
Kaiming He, Xiangyu Zhang |
Introduces Residual or Skip Connections to allow increase in the depth of a DNN |
link |
3 |
Evaluation of neural network architectures for embedded systems |
Read |
CNNs, CV , Image |
Comparison |
2017 |
IEEE ISCAS |
Adam Paszke, Alfredo Canziani, Eugenio Culurciello |
Compare CNN classification architectures on accuracy, memory footprint, parameters, operations count, inference time and power consumption. |
link |
4 |
SqueezeNet |
Read |
CNNs, CV , Image |
Architecture, Optimization-No. of params |
2016 |
arXiv |
Forrest N. Iandola, Song Han |
Explores model compression by using 1x1 convolutions called fire modules. |
link |
5 |
Attention is All you Need |
Read |
Attention, Text , Transformers |
Architecture |
2017 |
NIPS |
Ashish Vaswani, Illia Polosukhin, Noam Shazeer, Łukasz Kaiser |
Talks about Transformer architecture which brings SOTA performance for different tasks in NLP |
link |
6 |
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding |
Read |
Attention, Text , Transformers |
Embeddings |
2018 |
NAACL |
Jacob Devlin, Kenton Lee, Kristina Toutanova, Ming-Wei Chang |
BERT is an extension to Transformer based architecture which introduces a masked word pretraining and next sentence prediction task to pretrain the model for a wide variety of tasks. |
link |
7 |
Reformer: The Efficient Transformer |
Read |
Attention, Text , Transformers |
Architecture, Optimization-Memory, Optimization-No. of params |
2020 |
arXiv |
Anselm Levskaya, Lukasz Kaiser, Nikita Kitaev |
Overcome time and memory complexity of Transformers by bucketing Query, Keys and using Reversible residual connections. |
link |
8 |
Bag of Tricks for Image Classification with Convolutional Neural Networks |
Read |
CV , Image |
Optimizations, Tips & Tricks |
2018 |
arXiv |
Tong He, Zhi Zhang |
Shows a dozen tricks (mixup, label smoothing, etc.) to improve CNN accuracy and training time. |
link |
9 |
The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks |
Read |
NN Initialization, NNs |
Optimization-No. of params, Tips & Tricks |
2019 |
ICLR |
Jonathan Frankle, Michael Carbin |
Lottery ticket hypothesis: dense, randomly-initialized, feed-forward networks contain subnetworks (winning tickets) that—when trained in isolation— reach test accuracy comparable to the original network in a similar number of iterations. |
link |
10 |
Pix2Pix: Image-to-Image Translation with Conditional Adversarial Nets |
Read |
GANs, Image |
|
2017 |
CVPR |
Alexei A. Efros, Jun-Yan Zhu, Phillip Isola, Tinghui Zhou |
Image to image translation using Conditional GANs and dataset of image pairs from one domain to another. |
link |
11 |
Language-Agnostic BERT Sentence Embedding |
Read |
Attention, Siamese Network, Text , Transformers |
Embeddings |
2020 |
arXiv |
Fangxiaoyu Feng, Yinfei Yang |
A BERT model with multilingual sentence embeddings learned over 112 languages and Zero-shot learning over unseen languages. |
link |
12 |
T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer |
Read |
Attention, Text , Transformers |
|
2020 |
JMLR |
Colin Raffel, Noam Shazeer, Peter J. Liu, Wei Liu, Yanqi Zhou |
Presents a Text-to-Text transformer model with multi-task learning capabilities, simultaneously solving problems such as machine translation, document summarization, question answering, and classification tasks. |
link |
13 |
Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask |
Read |
NN Initialization, NNs |
Comparison, Optimization-No. of params, Tips & Tricks |
2019 |
NeurIPS |
Hattie Zhou, Janice Lan, Jason Yosinski, Rosanne Liu |
Follow up on Lottery Ticket Hypothesis exploring the effects of different Masking criteria as well as Mask-1 and Mask-0 actions. |
link |
14 |
SpanBERT: Improving Pre-training by Representing and Predicting Spans |
Read |
Question-Answering, Text , Transformers |
Pre-Training |
2020 |
TACL |
Danqi Chen, Mandar Joshi |
A different pre-training strategy for BERT model to improve performance for Question Answering task. |
link |
15 |
Learning to Extract Attribute Value from Product via Question Answering: A Multi-task Approach |
Read |
Question-Answering, Text , Transformers |
Zero-shot-learning |
2020 |
KDD |
Li Yang, Qifan Wang |
Question Answering BERT model used to extract attributes from products. Introduce further No Answer loss and distillation to promote zero shot learning. |
link |
16 |
VL-T5: Unifying Vision-and-Language Tasks via Text Generation |
Read |
CNNs, CV , Generative, Image , Large-Language-Models, Question-Answering, Text , Transformers |
Architecture, Embeddings, Multimodal, Pre-Training |
2021 |
arXiv |
Hao Tan, Jaemin Cho, Jie Le, Mohit Bansal |
Unifying two modalities (image and text) together in a single transformer model to solve multiple tasks in a single architecture using text prefixes similar to T5. |
link |