LinkNet: Exploiting Encoder Representations for Efficient Semantic Segmentation

June 14, 2017 ยท Entered Twilight ยท ๐Ÿ› Visual Communications and Image Processing

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 8.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, README.md, data, main.lua, misc, models, opts.lua, test.lua, train.lua

Authors Abhishek Chaurasia, Eugenio Culurciello arXiv ID 1707.03718 Category cs.CV: Computer Vision Cross-listed cs.LG Citations 1.6K Venue Visual Communications and Image Processing Repository https://github.com/e-lab/LinkNet โญ 174 Last Checked 1 month ago
Abstract
Pixel-wise semantic segmentation for visual scene understanding not only needs to be accurate, but also efficient in order to find any use in real-time application. Existing algorithms even though are accurate but they do not focus on utilizing the parameters of neural network efficiently. As a result they are huge in terms of parameters and number of operations; hence slow too. In this paper, we propose a novel deep neural network architecture which allows it to learn without any significant increase in number of parameters. Our network uses only 11.5 million parameters and 21.2 GFLOPs for processing an image of resolution 3x640x360. It gives state-of-the-art performance on CamVid and comparable results on Cityscapes dataset. We also compare our networks processing time on NVIDIA GPU and embedded system device with existing state-of-the-art architectures for different image resolutions.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision