scispace - formally typeset
Search or ask a question
Conference

Visual Communications and Image Processing 

About: Visual Communications and Image Processing is an academic conference. The conference publishes majorly in the area(s): Image processing & Motion estimation. Over the lifetime, 5000 publications have been published by the conference receiving 45761 citations.


Papers
More filters
Proceedings ArticleDOI
14 Jun 2017
TL;DR: In this paper, the authors proposed a novel deep neural network architecture which allows it to learn without any significant increase in number of parameters and achieves state-of-the-art performance on CamVid and Cityscapes dataset.
Abstract: Pixel-wise semantic segmentation for visual scene understanding not only needs to be accurate, but also efficient in order to find any use in real-time application. Existing algorithms even though are accurate but they do not focus on utilizing the parameters of neural network efficiently. As a result they are huge in terms of parameters and number of operations; hence slow too. In this paper, we propose a novel deep neural network architecture which allows it to learn without any significant increase in number of parameters. Our network uses only 11.5 million parameters and 21.2 GFLOPs for processing an image of resolution 3 × 640 × 360. It gives state-of-the-art performance on CamVid and comparable results on Cityscapes dataset. We also compare our networks processing time on NVIDIA GPU and embedded system device with existing state-of-the-art architectures for different image resolutions.

1,015 citations

Proceedings ArticleDOI
18 Jan 2004
TL;DR: This paper compares various background subtraction algorithms for detecting moving vehicles and pedestrians in urban traffic video sequences, considering approaches varying from simple techniques such as frame differencing and adaptive median filtering, to more sophisticated probabilistic modeling techniques.
Abstract: Identifying moving objects from a video sequence is a fundamental and critical task in many computer-vision applications. A common approach is to perform background subtraction, which identifies moving objects from the portion of a video frame that differs significantly from a background model. There are many challenges in developing a good background subtraction algorithm. First, it must be robust against changes in illumination. Second, it should avoid detecting non-stationary background objects such as swinging leaves, rain, snow, and shadow cast by moving objects. Finally, its internal background model should react quickly to changes in background such as starting and stopping of vehicles. In this paper, we compare various background subtraction algorithms for detecting moving vehicles and pedestrians in urban traffic video sequences. We consider approaches varying from simple techniques such as frame differencing and adaptive median filtering, to more sophisticated probabilistic modeling techniques. While complicated techniques often produce superior performance, our experiments show that simple techniques such as adaptive median filtering can produce good results with much lower computational complexity.

794 citations

Proceedings ArticleDOI
01 Feb 1990
TL;DR: In this article, a small population approach (coined as Micro-Genetic Algorithms--μGA) with some very simple genetic parameters was explored and it was shown that,μGA implementation reaches the near-optimal region much earlier than the SGA implementation.
Abstract: Simple Genetic Algorithms (SGA) have been shown to be useful tools for many function optimization problems. One present drawback of SGA is the time penalty involved in evaluating the fitness functions (performance indices) for large populations, generation after generation. This paper explores a small population approach (coined as Micro-Genetic Algorithms--μGA) with some very simple genetic parameters. It is shown that ,μGA implementation reaches the near-optimal region much earlier than the SGA implementation. The superior performance of the ,μGA in the presence of multimodality and their merits in solving non-stationary function optimization problems are demonstrated.

736 citations

Proceedings ArticleDOI
30 May 2000
TL;DR: The continuum between images and geometry used in image-based rendering techniques suggests that image- based rendering with traditional 3D graphics can be united in a joint image and geometry space.
Abstract: In this paper, we survey the techniques for image-based rendering. Unlike traditional 3D computer graphics in which 3D geometry of the scene is known, image-based rendering techniques render novel views directly from input images. Previous image-based rendering techniques can be classified into three categories according to how much geometric information is used: rendering without geometry, rendering with implicit geometry (i.e., correspondence), and rendering with explicit geometry (either with approximate or accurate geometry). We discuss the characteristics of these categories and their representative methods. The continuum between images and geometry used in image-based rendering techniques suggests that image-based rendering with traditional 3D graphics can be united in a joint image and geometry space.

516 citations

Proceedings ArticleDOI
07 Jan 2004
TL;DR: This work proposes a transformdomain Wyner-Ziv coding scheme for motion video that uses intraframe encoding, but interframe decoding, and shows significant gains above DCT-based intraframe coding and improvements over the pixel-domain Wynev video coder.
Abstract: In current interframe video compression systems, the encoder performs predictive coding to exploit the similarities of successive frames. The Wyner-Ziv Theorem on source coding with side information available only at the decoder suggests that an asymmetric video codec, where individual frames are encoded separately, but decoded conditionally (given temporally adjacent frames) could achieve similar efficiency. We propose a transformdomain Wyner-Ziv coding scheme for motion video that uses intraframe encoding, but interframe decoding. In this system, the transform coefficients of a Wyner-Ziv frame are encoded independently using a scalar quantizer and turbo coder. The decoder uses previously reconstructed frames to generate side information to conditionally decode the Wyner-Ziv frames. Simulation results show significant gains above DCT-based intraframe coding and improvements over the pixel-domain Wyner-Ziv video coder.

469 citations

Performance
Metrics
No. of papers from the Conference in previous years
YearPapers
2022112
20211
2020137
2019134
2018123
2017142