scispace - formally typeset
Search or ask a question
Author

Gian Luca Foresti

Bio: Gian Luca Foresti is an academic researcher from University of Udine. The author has contributed to research in topics: Object detection & Video tracking. The author has an hindex of 41, co-authored 364 publications receiving 6862 citations. Previous affiliations of Gian Luca Foresti include Sapienza University of Rome & University of Genoa.


Papers
More filters
Journal ArticleDOI
TL;DR: The proposed work addresses anomaly detection by means of trajectory analysis, an approach with several application fields, most notably video surveillance and traffic monitoring, based on single-class support vector machine (SVM) clustering, where the novelty detection SVM capabilities are used for the identification of anomalous trajectories.
Abstract: During the last years, the task of automatic event analysis in video sequences has gained an increasing attention among the research community. The application domains are disparate, ranging from video surveillance to automatic video annotation for sport videos or TV shots. Whatever the application field, most of the works in event analysis are based on two main approaches: the former based on explicit event recognition, focused on finding high-level, semantic interpretations of video sequences, and the latter based on anomaly detection. This paper deals with the second approach, where the final goal is not the explicit labeling of recognized events, but the detection of anomalous events differing from typical patterns. In particular, the proposed work addresses anomaly detection by means of trajectory analysis, an approach with several application fields, most notably video surveillance and traffic monitoring. The proposed approach is based on single-class support vector machine (SVM) clustering, where the novelty detection SVM capabilities are used for the identification of anomalous trajectories. Particular attention is given to trajectory classification in absence of a priori information on the distribution of outliers. Experimental results prove the validity of the proposed approach.

507 citations

Journal ArticleDOI
TL;DR: A trajectory clustering algorithm suited for video surveillance systems that clusters are organized in a tree-like structure that can be used to perform behaviour analysis, since it allows the identification of anomalous events.

267 citations

Journal ArticleDOI
01 Jan 2005
TL;DR: Ambient intelligence is a user-centric paradigm that supports a variety of artificial intelligence methods and works pervasively, nonintrusively, and transparently to aid the user.
Abstract: Ambient intelligence (AmI) is a new multidisciplinary paradigm rooted in the ideas of NormanAuthor of the Invisible Computer [32]. and Ubiquitous Computing. AmI fosters novel anthropomorphic human–machine models of interaction. In AmI, technologies are deployed to make computers disappear in the background, while the human user moves into the foreground in complete control of the augmented environment. AmI is a user-centric paradigm, it supports a variety of artificial intelligence methods and works pervasively, nonintrusively, and transparently to aid the user. AmI supports and promotes interdisciplinary research encompassing the technological, scientific and artistic fields creating a virtual support for embedded and distributed intelligence.

233 citations

Book ChapterDOI
Matej Kristan1, Ales Leonardis2, Jiří Matas3, Michael Felsberg4, Roman Pflugfelder5, Roman Pflugfelder6, Joni-Kristian Kamarainen, Martin Danelljan7, Luka Čehovin Zajc1, Alan Lukežič1, Ondrej Drbohlav3, Linbo He4, Yushan Zhang4, Yushan Zhang8, Song Yan, Jinyu Yang2, Gustavo Fernandez5, Alexander G. Hauptmann9, Alireza Memarmoghadam10, Alvaro Garcia-Martin11, Andreas Robinson4, Anton Varfolomieiev12, Awet Haileslassie Gebrehiwot11, Bedirhan Uzun13, Bin Yan14, Bing Li15, Chen Qian, Chi-Yi Tsai16, Christian Micheloni17, Dong Wang14, Fei Wang, Fei Xie18, Felix Järemo Lawin4, Fredrik K. Gustafsson19, Gian Luca Foresti17, Goutam Bhat7, Guangqi Chen, Haibin Ling20, Haitao Zhang, Hakan Cevikalp13, Haojie Zhao14, Haoran Bai21, Hari Chandana Kuchibhotla22, Hasan Saribas, Heng Fan20, Hossein Ghanei-Yakhdan23, Houqiang Li24, Houwen Peng25, Huchuan Lu14, Hui Li26, Javad Khaghani27, Jesús Bescós11, Jianhua Li14, Jianlong Fu25, Jiaqian Yu28, Jingtao Xu28, Josef Kittler29, Jun Yin, Junhyun Lee30, Kaicheng Yu31, Kaiwen Liu15, Kang Yang32, Kenan Dai14, Li Cheng27, Li Zhang33, Lijun Wang14, Linyuan Wang, Luc Van Gool7, Luca Bertinetto, Matteo Dunnhofer17, Miao Cheng, Mohana Murali Dasari22, Ning Wang32, Pengyu Zhang14, Philip H. S. Torr33, Qiang Wang, Radu Timofte7, Rama Krishna Sai Subrahmanyam Gorthi22, Seokeon Choi34, Seyed Mojtaba Marvasti-Zadeh27, Shaochuan Zhao26, Shohreh Kasaei35, Shoumeng Qiu15, Shuhao Chen14, Thomas B. Schön19, Tianyang Xu29, Wei Lu, Weiming Hu15, Wengang Zhou24, Xi Qiu, Xiao Ke36, Xiaojun Wu26, Xiaolin Zhang15, Xiaoyun Yang, Xue-Feng Zhu26, Yingjie Jiang26, Yingming Wang14, Yiwei Chen28, Yu Ye36, Yuezhou Li36, Yuncon Yao18, Yunsung Lee30, Yuzhang Gu15, Zezhou Wang14, Zhangyong Tang26, Zhen-Hua Feng29, Zhijun Mai37, Zhipeng Zhang15, Zhirong Wu25, Ziang Ma 
23 Aug 2020
TL;DR: A significant novelty is introduction of a new VOT short-term tracking evaluation methodology, and introduction of segmentation ground truth in the VOT-ST2020 challenge – bounding boxes will no longer be used in theVDT challenges.
Abstract: The Visual Object Tracking challenge VOT2020 is the eighth annual tracker benchmarking activity organized by the VOT initiative. Results of 58 trackers are presented; many are state-of-the-art trackers published at major computer vision conferences or in journals in the recent years. The VOT2020 challenge was composed of five sub-challenges focusing on different tracking domains: (i) VOT-ST2020 challenge focused on short-term tracking in RGB, (ii) VOT-RT2020 challenge focused on “real-time” short-term tracking in RGB, (iii) VOT-LT2020 focused on long-term tracking namely coping with target disappearance and reappearance, (iv) VOT-RGBT2020 challenge focused on short-term tracking in RGB and thermal imagery and (v) VOT-RGBD2020 challenge focused on long-term tracking in RGB and depth imagery. Only the VOT-ST2020 datasets were refreshed. A significant novelty is introduction of a new VOT short-term tracking evaluation methodology, and introduction of segmentation ground truth in the VOT-ST2020 challenge – bounding boxes will no longer be used in the VOT-ST challenges. A new VOT Python toolkit that implements all these novelites was introduced. Performance of the tested trackers typically by far exceeds standard baselines. The source code for most of the trackers is publicly available from the VOT page. The dataset, the evaluation kit and the results are publicly available at the challenge website (http://votchallenge.net).

158 citations

Proceedings ArticleDOI
12 Mar 2018
TL;DR: In this paper, a slice convolution block is introduced to capture vertical food traits that are common to a large number of categories (i.e., 15% of the whole data in current datasets).
Abstract: Image-based food recognition pose new challenges for mainstream computer vision algorithms. Recent works in the field focused either on hand-crafted representations or on learning these by exploiting deep neural networks (DNN). Despite the success of DNN-based works, these exploit off-the-shelf deep architectures which are not cast to the specific food classification problem. We believe that better results can be obtained if the architecture is defined with respect to an analysis of the food composition. Following such an intuition, this work introduces a new deep scheme that is designed to handle the food structure. In particular, we focus on the vertical food traits that are common to a large number of categories (i.e., 15% of the whole data in current datasets). Towards the final objective, we first introduce a slice convolution block to capture such specific information. Then, we leverage on the recent success of deep residual blocks and combine those with the sliced convolution to produce the classification score. Extensive evaluations on three benchmark datasets demonstrated that our solution has better performance than existing approaches (e.g., a top–1 accuracy of 90.27% on the Food-101 dataset).

151 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis.
Abstract: Machine Learning is the study of methods for programming computers to learn. Computers are applied to a wide range of tasks, and for most of these it is relatively easy for programmers to design and implement the necessary software. However, there are many tasks for which this is difficult or impossible. These can be divided into four general categories. First, there are problems for which there exist no human experts. For example, in modern automated manufacturing facilities, there is a need to predict machine failures before they occur by analyzing sensor readings. Because the machines are new, there are no human experts who can be interviewed by a programmer to provide the knowledge necessary to build a computer system. A machine learning system can study recorded data and subsequent machine failures and learn prediction rules. Second, there are problems where human experts exist, but where they are unable to explain their expertise. This is the case in many perceptual tasks, such as speech recognition, hand-writing recognition, and natural language understanding. Virtually all humans exhibit expert-level abilities on these tasks, but none of them can describe the detailed steps that they follow as they perform them. Fortunately, humans can provide machines with examples of the inputs and correct outputs for these tasks, so machine learning algorithms can learn to map the inputs to the outputs. Third, there are problems where phenomena are changing rapidly. In finance, for example, people would like to predict the future behavior of the stock market, of consumer purchases, or of exchange rates. These behaviors change frequently, so that even if a programmer could construct a good predictive computer program, it would need to be rewritten frequently. A learning program can relieve the programmer of this burden by constantly modifying and tuning a set of learned prediction rules. Fourth, there are applications that need to be customized for each computer user separately. Consider, for example, a program to filter unwanted electronic mail messages. Different users will need different filters. It is unreasonable to expect each user to program his or her own rules, and it is infeasible to provide every user with a software engineer to keep the rules up-to-date. A machine learning system can learn which mail messages the user rejects and maintain the filtering rules automatically. Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis. Statistics focuses on understanding the phenomena that have generated the data, often with the goal of testing different hypotheses about those phenomena. Data mining seeks to find patterns in the data that are understandable by people. Psychological studies of human learning aspire to understand the mechanisms underlying the various learning behaviors exhibited by people (concept learning, skill acquisition, strategy change, etc.).

13,246 citations

Christopher M. Bishop1
01 Jan 2006
TL;DR: Probability distributions of linear models for regression and classification are given in this article, along with a discussion of combining models and combining models in the context of machine learning and classification.
Abstract: Probability Distributions.- Linear Models for Regression.- Linear Models for Classification.- Neural Networks.- Kernel Methods.- Sparse Kernel Machines.- Graphical Models.- Mixture Models and EM.- Approximate Inference.- Sampling Methods.- Continuous Latent Variables.- Sequential Data.- Combining Models.

10,141 citations

01 Jan 2004
TL;DR: Comprehensive and up-to-date, this book includes essential topics that either reflect practical significance or are of theoretical importance and describes numerous important application areas such as image based rendering and digital libraries.
Abstract: From the Publisher: The accessible presentation of this book gives both a general view of the entire computer vision enterprise and also offers sufficient detail to be able to build useful applications. Users learn techniques that have proven to be useful by first-hand experience and a wide range of mathematical methods. A CD-ROM with every copy of the text contains source code for programming practice, color images, and illustrative movies. Comprehensive and up-to-date, this book includes essential topics that either reflect practical significance or are of theoretical importance. Topics are discussed in substantial and increasing depth. Application surveys describe numerous important application areas such as image based rendering and digital libraries. Many important algorithms broken down and illustrated in pseudo code. Appropriate for use by engineers as a comprehensive reference to the computer vision enterprise.

3,627 citations

01 Jan 2003

3,093 citations

Journal ArticleDOI
TL;DR: This article has reviewed the reasons why people want to love or leave the venerable (but perhaps hoary) MSE and reviewed emerging alternative signal fidelity measures and discussed their potential application to a wide variety of problems.
Abstract: In this article, we have reviewed the reasons why we (collectively) want to love or leave the venerable (but perhaps hoary) MSE. We have also reviewed emerging alternative signal fidelity measures and discussed their potential application to a wide variety of problems. The message we are trying to send here is not that one should abandon use of the MSE nor to blindly switch to any other particular signal fidelity measure. Rather, we hope to make the point that there are powerful, easy-to-use, and easy-to-understand alternatives that might be deployed depending on the application environment and needs. While we expect (and indeed, hope) that the MSE will continue to be widely used as a signal fidelity measure, it is our greater desire to see more advanced signal fidelity measures being used, especially in applications where perceptual criteria might be relevant. Ideally, the performance of a new signal processing algorithm might be compared to other algorithms using several fidelity criteria. Lastly, we hope that we have given further motivation to the community to consider recent advanced signal fidelity measures as design criteria for optimizing signal processing algorithms and systems. It is in this direction that we believe that the greatest benefit eventually lies.

2,601 citations