Optimal subsampling for quantile regression in big data
Citations
36 citations
21 citations
18 citations
Cites background from "Optimal subsampling for quantile re..."
...17 The asymptotic mean squared error (AMSE) of ?̃? is equal to the trace of Σ, which is given by AMSE(?̃?) = tr(Σ), (10)...
[...]
18 citations
Cites background from "Optimal subsampling for quantile re..."
...Wang & Ma (2020) considered the optimal subsampling for quantile regression in big data.25 For the above-mentioned subsampling methods, one common assumption is that the data is stored in one location....
[...]
9 citations
Cites result from "Optimal subsampling for quantile re..."
...In the process of revising our manuscript, we have noticed that Wang & Ma (2020) and Ai et al. (2020a) obtain similar results on optimal subsampling for QR in the context of big data....
[...]
References
4,862 citations
2,686 citations
"Optimal subsampling for quantile re..." refers background or methods in this paper
...However, the interior point algorithm still need polynomial time for optimization; its worst-case time complexity is O(N5/2p3), where N is the sample size and p is the dimension of the regression coefficient (Sec 6.4.4 of Koenker, 2005)....
[...]
...Whilst for linear median regression, under some conditions, the overall time complexity is O(N1+ap3 log n), where 0 a 0.5 (Theorem 6.3 of Koenker, 2005)....
[...]
...For the first term on the right hand side of (S.6), following an approach similar to that in Section 4.2 of Koenker (2005) under the conditions in Assumption 1, we have n N N∑ i=1 E(Z2ni) = n N N∑ i=1 ∫ vi 0 {Fε|X(s, xi)− Fε|X(0, xi)}ds = √ n N N∑ i=1 ∫ λTxi 0 {Fε|X(t/ √ n, xi)− Fε|X(0, xi)}dt =…...
[...]
...(5) As shown in Theorem 4.1 of Koenker (2005), under Assumption 1, the full data estimator β̂ satisfies that {τ(1− τ)D−1N DN0D−1N }−1/2 √ N(β̂ − βt) −→ N(0, I), (6) in distribution, where N(0, I) represents a multivariate standard normal distribution, and βt stands for the true value of β....
[...]
...Here we adopt the set of regularity conditions used in Koenker (2005) and list them below as Assumption 1 for completeness....
[...]
1,076 citations
"Optimal subsampling for quantile re..." refers background or methods in this paper
...ce also has an optimality interpretation in terms of optimal experimental design; it is termed the L-optimality criterion, where \L" stands for \linear transformation" of the estimator (see Atkinson et al., 2007). Using this criterion we are able to obtain the explicit expression of optimal subsampling probabilities in the following theorem. Theorem 2 (L-optimality) If the sampling probabilities ˇ i, i= 1;:::...
[...]
...This choice also has an optimality interpretation in terms of optimal experimental design; it is termed the L-optimality criterion, where “L” stands for “linear transformation” of the estimator (see Atkinson et al., 2007)....
[...]
...ng probabilities that minimize the asymptotic MSE of 1e, that is, the ˇ i’s that minimize the trace of n D 1 N V ˇD 1 N . This is called the A-optimality criterion in optimal experimental design (see Atkinson et al., 2007). Theorem 3 (A-optimality) If the sampling probabilities ˇ i, i= 1;:::;Nare chosen as ˇAopt i = j˝ I(" i<0)jkD 1 N xk P N j=1 j˝ I(" j <0)jkD 1 N x jk ;i= 1;2;:::;N; then the total asy...
[...]
...This is called the A-optimality criterion in optimal experimental design (see Atkinson et al., 2007)....
[...]
897 citations
"Optimal subsampling for quantile re..." refers background in this paper
...In big data problems, because data are often collected from different sources with different times and locations, the homoscedasticity assumption is often not valid (Fan et al., 2014), which makes quantile regression a natural candidate as an analysis tool....
[...]
733 citations