Top Cited Papers: CVPR 2015

Longuet-Higgins Prize(Test-of-Time)

2015“Histograms of oriented gradients for human detection”N. Dalal, B. Triggs

Best Paper Award

2015“DynamicFusion: Reconstruction and Tracking of Non-rigid Scenes in Real-Time”R. A. Newcombe, D. Fox, S. M. Seitz

Honorable Mention

2015“Efficient Globally Optimal Consensus Maximisation with Tree Search”T.-J. Chin, P. Purkait, A. Eriksson, D. Suter
2015“Fully Convolutional Networks for Semantic Segmentation”J. Long, E. Shelhamer, T. Darrell
2015“Picture: A Probabilistic Programming Language for Scene Perception”T. D. Kulkarni, P. Kohli, J. B. Tenenbaum, V. Mansinghka

Best Student Paper Award

2015“Category-Specific Object Reconstruction from a Single Image”A. Kar, S. Tulsiani, J. Carreira, J. Malik

Curated Papers:

Going deeper with convolutions

We propose a deep convolutional neural network architecture codenamed Inception that achieves the new state of the art for classification and detection in the ImageNet Large-Scale Visual Recognition Challenge 2014 (ILSVRC14). The main hallmark of this architecture is the improved utilization of the computing resources inside the network. By a carefully crafted design, we increased the depth and width of the network while keeping the computational budget constant. To optimize quality, the architectural decisions were based on the Hebbian principle and the intuition of multi-scale processing. One particular incarnation used in our submission for ILSVRC14 is called GoogLeNet, a 22 layers deep network, the quality of which is assessed in the context of classification and detection.

Single image super-resolution from transformed self-exemplars

Self-similarity based super-resolution (SR) algorithms are able to produce visually pleasing results without extensive training on external databases. Such algorithms exploit the statistical prior that patches in a natural image tend to recur within and across scales of the same image. However, the internal dictionary obtained from the given image may not always be sufficiently expressive to cover the textural appearance variations in the scene. In this paper, we extend self-similarity based SR to overcome this drawback. We expand the internal patch search space by allowing geometric variations. We do so by explicitly localizing planes in the scene and using the detected perspective geometry to guide the patch search process. We also incorporate additional affine transformations to accommodate local shape variations. We propose a compositional model to simultaneously handle both types of transformations. We extensively evaluate the performance in both urban and natural scenes. Even without using any external training databases, we achieve significantly superior results on urban scenes, while maintaining comparable performance on natural scenes as other state-of-the-art SR algorithms.

Learning to compare image patches via convolutional neural networks

In this paper we show how to learn directly from image data (i.e., without resorting to manually-designed features) a general similarity function for comparing image patches, which is a task of fundamental importance for many computer vision problems. To encode such a function, we opt for a CNN-based model that is trained to account for a wide variety of changes in image appearance. To that end, we explore and study multiple neural network architectures, which are specifically adapted to this task. We show that such an approach can significantly outperform the state-of-the-art on several problems and benchmark datasets.

Full List:

  1. Going deeper with convolutions Christian Szegedy;Wei Liu;Yangqing Jia;Pierre Sermanet;Scott Reed;Dragomir Anguelov;Dumitru Erhan;Vincent Vanhoucke;Andrew Rabinovich Publication Year: 2015,Page(s):1 – 9 Cited by: Papers (10368) | Patents (41)
  2. Fully convolutional networks for semantic segmentation Jonathan Long;Evan Shelhamer;Trevor Darrell Publication Year: 2015,Page(s):3431 – 3440 Cited by: Papers (7539) | Patents (95)
  3. FaceNet: A unified embedding for face recognition and clustering Florian Schroff;Dmitry Kalenichenko;James Philbin Publication Year: 2015,Page(s):815 – 823 Cited by: Papers (2780) | Patents (66)
  4. Long-term recurrent convolutional networks for visual recognition and description Jeff Donahue;Lisa Anne Hendricks;Sergio Guadarrama;Marcus Rohrbach;Subhashini Venugopalan;Trevor Darrell;Kate Saenko Publication Year: 2015,Page(s):2625 – 2634 Cited by: Papers (1332) | Patents (24)
  5. Show and tell: A neural image caption generator Oriol Vinyals;Alexander Toshev;Samy Bengio;Dumitru Erhan Publication Year: 2015,Page(s):3156 – 3164 Cited by: Papers (1314) | Patents (15)
  6. Deep visual-semantic alignments for generating image descriptions Andrej Karpathy;Li Fei-Fei Publication Year: 2015,Page(s):3128 – 3137 Cited by: Papers (1166) | Patents (19)
  7. Person re-identification by Local Maximal Occurrence representation and metric learning Shengcai Liao;Yang Hu;Xiangyu Zhu;Stan Z. Li Publication Year: 2015,Page(s):2197 – 2206 Cited by: Papers (850) | Patents (2)
  8. Single image super-resolution from transformed self-exemplars Jia-Bin Huang;Abhishek Singh;Narendra Ahuja Publication Year: 2015,Page(s):5197 – 5206 Cited by: Papers (626) | Patents (3)
  9. Deep neural networks are easily fooled: High confidence predictions for unrecognizable images Anh Nguyen;Jason Yosinski;Jeff Clune Publication Year: 2015,Page(s):427 – 436 Cited by: Papers (576) | Patents (4)
  10. An improved deep learning architecture for person re-identification Ejaz Ahmed;Michael Jones;Tim K. Marks Publication Year: 2015,Page(s):3908 – 3916 Cited by: Papers (560) | Patents (2)
  11. Object scene flow for autonomous vehicles Moritz Menze;Andreas Geiger Publication Year: 2015,Page(s):3061 – 3070 Cited by: Papers (547) | Patents (2)
  12. CIDEr: Consensus-based image description evaluation Ramakrishna Vedantam;C. Lawrence Zitnick;Devi Parikh Publication Year: 2015,Page(s):4566 – 4575 Cited by: Papers (534) | Patents (4)
  13. Action recognition with trajectory-pooled deep-convolutional descriptors Limin Wang;Yu Qiao;Xiaoou Tang Publication Year: 2015,Page(s):4305 – 4314 Cited by: Papers (519) | Patents (2)
  14. Learning to compare image patches via convolutional neural networks Sergey Zagoruyko;Nikos Komodakis Publication Year: 2015,Page(s):4353 – 4361 Cited by: Papers (517) | Patents (13)
  15. A convolutional neural network cascade for face detection Haoxiang Li;Zhe Lin;Xiaohui Shen;Jonathan Brandt;Gang Hua Publication Year: 2015,Page(s):5325 – 5334 Cited by: Papers (497) | Patents (5)
  16. Hypercolumns for object segmentation and fine-grained localization Bharath Hariharan;Pablo Arbeláez;Ross Girshick;Jitendra Malik Publication Year: 2015,Page(s):447 – 456 Cited by: Papers (491) | Patents (1)
  17. Long-term correlation tracking Chao Ma;Xiaokang Yang;Chongyang Zhang;Ming-Hsuan Yang Publication Year: 2015,Page(s):5388 – 5396 Cited by: Papers (455)
  18. Supervised Discrete Hashing Fumin Shen;Chunhua Shen;Wei Liu;Heng Tao Shen Publication Year: 2015,Page(s):37 – 45 Cited by: Papers (427)
  19. Understanding deep image representations by inverting them Aravindh Mahendran;Andrea Vedaldi Publication Year: 2015,Page(s):5188 – 5196 Cited by: Papers (421) | Patents (4)
  20. From captions to visual concepts and back Hao Fang;Saurabh Gupta;Forrest Iandola;Rupesh K. Srivastava;Li Deng;Piotr Dollár;Jianfeng Gao;Xiaodong He;Margaret Mitchell;John C. Platt;C. Lawrence Zitnick;Geoffrey Zweig Publication Year: 2015,Page(s):1473 – 1482 Cited by: Papers (416) | Patents (8)
  21. Saliency detection by multi-context deep learning Rui Zhao;Wanli Ouyang;Hongsheng Li;Xiaogang Wang Publication Year: 2015,Page(s):1265 – 1274 Cited by: Papers (400) | Patents (3)
  22. Beyond short snippets: Deep networks for video classification Joe Yue-Hei Ng;Matthew Hausknecht;Sudheendra Vijayanarasimhan;Oriol Vinyals;Rajat Monga;George Toderici Publication Year: 2015,Page(s):4694 – 4702 Cited by: Papers (362) | Patents (2)
  23. SUN RGB-D: A RGB-D scene understanding benchmark suite Shuran Song;Samuel P. Lichtenberg;Jianxiong Xiao Publication Year: 2015,Page(s):567 – 576 Cited by: Papers (359) | Patents (5)
  24. Deeply learned face representations are sparse, selective, and robust Yi Sun;Xiaogang Wang;Xiaoou Tang Publication Year: 2015,Page(s):2892 – 2900 Cited by: Papers (354)
  25. EpicFlow: Edge-preserving interpolation of correspondences for optical flow Jerome Revaud;Philippe Weinzaepfel;Zaid Harchaoui;Cordelia Schmid Publication Year: 2015,Page(s):1164 – 1172 Cited by: Papers (342) | Patents (3)
  26. Is object localization for free? – Weakly-supervised learning with convolutional neural networks Maxime Oquab;Léon Bottou;Ivan Laptev;Josef Sivic Publication Year: 2015,Page(s):685 – 694 Cited by: Papers (341) | Patents (5)
  27. Cross-scene crowd counting via deep convolutional neural networks Cong Zhang;Hongsheng Li;Xiaogang Wang;Xiaokang Yang Publication Year: 2015,Page(s):833 – 841 Cited by: Papers (335) | Patents (4)
  28. ActivityNet: A large-scale video benchmark for human activity understanding Fabian Caba Heilbron;Victor Escorcia;Bernard Ghanem;Juan Carlos Niebles Publication Year: 2015,Page(s):961 – 970 Cited by: Papers (328)
  29. MUlti-Store Tracker (MUSTer): A cognitive psychology inspired approach to object tracking Zhibin Hong;Zhe Chen;Chaohui Wang;Xue Mei;Danil Prokhorov;Dacheng Tao Publication Year: 2015,Page(s):749 – 758 Cited by: Papers (324)
  30. Convolutional neural networks at constrained time cost Kaiming He;Jian Sun Publication Year: 2015,Page(s):5353 – 5360 Cited by: Papers (317)
  31. Deep convolutional neural fields for depth estimation from a single image Fayao Liu;Chunhua Shen;Guosheng Lin Publication Year: 2015,Page(s):5162 – 5170 Cited by: Papers (317) | Patents (1)
  32. Efficient object localization using Convolutional Networks Jonathan Tompson;Ross Goroshin;Arjun Jain;Yann LeCun;Christoph Bregler Publication Year: 2015,Page(s):648 – 656 Cited by: Papers (316)
  33. DynamicFusion: Reconstruction and tracking of non-rigid scenes in real-time Richard A. Newcombe;Dieter Fox;Steven M. Seitz Publication Year: 2015,Page(s):343 – 352 Cited by: Papers (315) | Patents (6)
  34. Simultaneous feature learning and hash coding with deep neural networks Hanjiang Lai;Yan Pan;Ye Liu;Shuicheng Yan Publication Year: 2015,Page(s):3270 – 3278 Cited by: Papers (315) | Patents (13)
  35. Evaluation of output embeddings for fine-grained image classification Zeynep Akata;Scott Reed;Daniel Walter;Honglak Lee;Bernt Schiele Publication Year: 2015,Page(s):2927 – 2936 Cited by: Papers (279)
  36. Deep networks for saliency detection via local estimation and global search Lijun Wang;Huchuan Lu;Xiang Ruan;Ming-Hsuan Yang Publication Year: 2015,Page(s):3183 – 3192 Cited by: Papers (277) | Patents (4)
  37. Recurrent convolutional neural network for object recognition Ming Liang;Xiaolin Hu Publication Year: 2015,Page(s):3367 – 3375 Cited by: Papers (268) | Patents (2)
  38. Fast and accurate image upscaling with super-resolution forests Samuel Schulter;Christian Leistner;Horst Bischof Publication Year: 2015,Page(s):3791 – 3799 Cited by: Papers (264) | Patents (3)
  39. A large-scale car dataset for fine-grained categorization and verification Linjie Yang;Ping Luo;Chen Change Loy;Xiaoou Tang Publication Year: 2015,Page(s):3973 – 3981 Cited by: Papers (262) | Patents (2)
  40. Deep filter banks for texture recognition and segmentation Mircea Cimpoi;Subhransu Maji;Andrea Vedaldi Publication Year: 2015,Page(s):3828 – 3836 Cited by: Papers (262) | Patents (2)
  41. Learning to generate chairs with convolutional neural networks Alexey Dosovitskiy;Jost Tobias Springenberg;Thomas Brox Publication Year: 2015,Page(s):1538 – 1546 Cited by: Papers (260) | Patents (3)
  42. Pushing the frontiers of unconstrained face detection and recognition: IARPA Janus Benchmark A Brendan F. Klare;Ben Klein;Emma Taborsky;Austin Blanton;Jordan Cheney;Kristen Allen;Patrick Grother;Alan Mah;Mark Burge;Anil K. Jain Publication Year: 2015,Page(s):1931 – 1939 Cited by: Papers (256) | Patents (1)
  43. Computing the stereo matching cost with a convolutional neural network Jure Žbontar;Yann LeCun Publication Year: 2015,Page(s):1592 – 1599 Cited by: Papers (245) | Patents (5)
  44. Finding action tubes Georgia Gkioxari;Jitendra Malik Publication Year: 2015,Page(s):759 – 768 Cited by: Papers (239) | Patents (5)
  45. Effective face frontalization in unconstrained images Tal Hassner;Shai Harel;Eran Paz;Roee Enbar Publication Year: 2015,Page(s):4295 – 4304 Cited by: Papers (235) | Patents (1)
  46. Learning a convolutional neural network for non-uniform motion blur removal Jian Sun;Wenfei Cao;Zongben Xu;Jean Ponce Publication Year: 2015,Page(s):769 – 777 Cited by: Papers (231) | Patents (1)
  47. Saliency-aware geodesic video object segmentation Wenguan Wang;Jianbing Shen;Fatih Porikli Publication Year: 2015,Page(s):3395 – 3402 Cited by: Papers (230)
  48. Learning to rank in person re-identification with metric ensembles Sakrapee Paisitkriangkrai;Chunhua Shen;Anton van den Hengel Publication Year: 2015,Page(s):1846 – 1855 Cited by: Papers (229)
  49. Accurate depth map estimation from a lenslet light field camera Hae-Gon Jeon;Jaesik Park;Gyeongmin Choe;Jinsun Park;Yunsu Bok;Yu-Wing Tai;In So Kweon Publication Year: 2015,Page(s):1547 – 1555 Cited by: Papers (212) | Patents (2)
  50. Filtered channel features for pedestrian detection Shanshan Zhang;Rodrigo Benenson;Bernt Schiele Publication Year: 2015,Page(s):1751 – 1760 Cited by: Papers (206) | Patents (1)

https://ieeexplore.ieee.org/xpl/conhome/7293313/proceeding?sortType=paper-citations&rowsPerPage=50&pageNumber=1

Leave a Reply Cancel reply