scispace - formally typeset
Search or ask a question
Institution

Netflix

CompanyLos Gatos, California, United States
About: Netflix is a company organization based out in Los Gatos, California, United States. It is known for research contribution in the topics: Video quality & Recommender system. The organization has 422 authors who have published 514 publications receiving 11838 citations. The organization is also known as: Netflix, Inc. & Netflix.com.


Papers
More filters
Proceedings Article
01 Jan 2016
TL;DR: In this article, the authors apply recurrent neural networks (RNN) on a new domain, namely recommender systems, and propose an RNN-based approach for session-based recommendations.
Abstract: We apply recurrent neural networks (RNN) on a new domain, namely recommender systems. Real-life recommender systems often face the problem of having to base recommendations only on short session-based data (e.g. a small sportsware website) instead of long user histories (as in the case of Netflix). In this situation the frequently praised matrix factorization approaches are not accurate. This problem is usually overcome in practice by resorting to item-to-item recommendations, i.e. recommending similar items. We argue that by modeling the whole session, more accurate recommendations can be provided. We therefore propose an RNN-based approach for session-based recommendations. Our approach also considers practical aspects of the task and introduces several modifications to classic RNNs such as a ranking loss function that make it more viable for this specific problem. Experimental results on two data-sets show marked improvements over widely used approaches.

989 citations

Posted Content
TL;DR: It is argued that by modeling the whole session, more accurate recommendations can be provided by an RNN-based approach for session-based recommendations, and introduced several modifications to classic RNNs such as a ranking loss function that make it more viable for this specific problem.
Abstract: We apply recurrent neural networks (RNN) on a new domain, namely recommender systems. Real-life recommender systems often face the problem of having to base recommendations only on short session-based data (e.g. a small sportsware website) instead of long user histories (as in the case of Netflix). In this situation the frequently praised matrix factorization approaches are not accurate. This problem is usually overcome in practice by resorting to item-to-item recommendations, i.e. recommending similar items. We argue that by modeling the whole session, more accurate recommendations can be provided. We therefore propose an RNN-based approach for session-based recommendations. Our approach also considers practical aspects of the task and introduces several modifications to classic RNNs such as a ranking loss function that make it more viable for this specific problem. Experimental results on two data-sets show marked improvements over widely used approaches.

950 citations

Proceedings ArticleDOI
17 Aug 2014
TL;DR: This work suggests an alternative approach: rather than presuming that capacity estimation is required, it is perhaps better to begin by using only the buffer, and then ask whencapacity estimation is needed, which allows us to reduce the rebuffer rate by 10-20% compared to Netflix's then-default ABR algorithm, while delivering a similar average video rate.
Abstract: Existing ABR algorithms face a significant challenge in estimating future capacity: capacity can vary widely over time, a phenomenon commonly observed in commercial services. In this work, we suggest an alternative approach: rather than presuming that capacity estimation is required, it is perhaps better to begin by using only the buffer, and then ask when capacity estimation is needed. We test the viability of this approach through a series of experiments spanning millions of real users in a commercial service. We start with a simple design which directly chooses the video rate based on the current buffer occupancy. Our own investigation reveals that capacity estimation is unnecessary in steady state; however using simple capacity estimation (based on immediate past throughput) is important during the startup phase, when the buffer itself is growing from empty. This approach allows us to reduce the rebuffer rate by 10-20% compared to Netflix's then-default ABR algorithm, while delivering a similar average video rate, and a higher video rate in steady state.

931 citations

Journal ArticleDOI
28 Dec 2015
TL;DR: The motivations behind and approach that Netflix uses to improve the recommendation algorithms are explained, combining A/B testing focused on improving member retention and medium term engagement, as well as offline experimentation using historical member engagement data.
Abstract: This article discusses the various algorithms that make up the Netflix recommender system, and describes its business purpose. We also describe the role of search and related algorithms, which for us turns into a recommendations problem as well. We explain the motivations behind and review the approach that we use to improve the recommendation algorithms, combining A/B testing focused on improving member retention and medium term engagement, as well as offline experimentation using historical member engagement data. We discuss some of the issues in designing and interpreting A/B tests. Finally, we describe some current areas of focused innovation, which include making our recommender system global and language aware.

906 citations

Proceedings ArticleDOI
23 Apr 2018
TL;DR: In this article, a variational autoencoder (VAE) was extended to collaborative filtering for implicit feedback, and a generative model with multinomial likelihood and Bayesian inference for parameter estimation was proposed.
Abstract: We extend variational autoencoders (VAEs) to collaborative filtering for implicit feedback. This non-linear probabilistic model enables us to go beyond the limited modeling capacity of linear factor models which still largely dominate collaborative filtering research.We introduce a generative model with multinomial likelihood and use Bayesian inference for parameter estimation. Despite widespread use in language modeling and economics, the multinomial likelihood receives less attention in the recommender systems literature. We introduce a different regularization parameter for the learning objective, which proves to be crucial for achieving competitive performance. Remarkably, there is an efficient way to tune the parameter using annealing. The resulting model and learning algorithm has information-theoretic connections to maximum entropy discrimination and the information bottleneck principle. Empirically, we show that the proposed approach significantly outperforms several state-of-the-art baselines, including two recently-proposed neural network approaches, on several real-world datasets. We also provide extended experiments comparing the multinomial likelihood with other commonly used likelihood functions in the latent factor collaborative filtering literature and show favorable results. Finally, we identify the pros and cons of employing a principled Bayesian inference approach and characterize settings where it provides the most significant improvements.

637 citations


Authors

Showing all 424 results

NameH-indexPapersCitations
Fei Sha5217615360
Nikos Vlassis451468839
Tony Jebara4517510569
Roberto Sanchis-Ojeda37836504
Ioannis Katsavounidis351103556
Anne Aaron26484933
Dong Liu26542497
Xavier Amatriain26903792
Linas Baltrunas24395085
Anush K. Moorthy23517168
Zhi Li22791643
Randall A. Lewis20461649
Lorin Hochstein20391192
Ioannis Papapanagiotou19631242
Arun Kejariwal19851219
Network Information
Related Institutions (5)
Microsoft
86.9K papers, 4.1M citations

82% related

Google
39.8K papers, 2.1M citations

82% related

Carnegie Mellon University
104.3K papers, 5.9M citations

81% related

Hewlett-Packard
59.8K papers, 1.4M citations

80% related

IBM
253.9K papers, 7.4M citations

79% related

Performance
Metrics
No. of papers from the Institution in previous years
YearPapers
20231
20221
202156
202047
201956
201875