scispace - formally typeset
Search or ask a question
Topic

Feasible region

About: Feasible region is a research topic. Over the lifetime, 4642 publications have been published within this topic receiving 101233 citations. The topic is also known as: feasible set & search space.


Papers
More filters
Journal ArticleDOI
TL;DR: GA's population-based approach and ability to make pair-wise comparison in tournament selection operator are exploited to devise a penalty function approach that does not require any penalty parameter to guide the search towards the constrained optimum.

3,495 citations

Book
01 Aug 1989
TL;DR: Mathematical Background Topics from Linear Algebra Single Objective Linear Programming Determining all Alternative Optima Comments about Objective Row Parametric Programming Utility Functions, Nondominated Criterion Vectors and Efficient Points Point Estimate Weighted-sums Approach.
Abstract: Mathematical Background Topics from Linear Algebra Single Objective Linear Programming Determining all Alternative Optima Comments about Objective Row Parametric Programming Utility Functions, Nondominated Criterion Vectors and Efficient Points Point Estimate Weighted-sums Approach Optimal Weighting Vectors, Scaling and Reduced Feasible Region Methods Vector-Maximum Algorithms Goal Programming Filtering and Set Discretization Multiple Objective Linear Fractional Programming Interactive Procedures Interactive Weighted Tchebycheff Procedure Tchebycheff/Weighted-Sums Implementation Applications Future Directions Index.

3,280 citations

Journal ArticleDOI
TL;DR: This paper considers regularized block multiconvex optimization, where the feasible set and objective function are generally nonconvex but convex in each block of variables and proposes a generalized block coordinate descent method.
Abstract: This paper considers regularized block multiconvex optimization, where the feasible set and objective function are generally nonconvex but convex in each block of variables. It also accepts nonconvex blocks and requires these blocks to be updated by proximal minimization. We review some interesting applications and propose a generalized block coordinate descent method. Under certain conditions, we show that any limit point satisfies the Nash equilibrium conditions. Furthermore, we establish global convergence and estimate the asymptotic convergence rate of the method by assuming a property based on the Kurdyka--Łojasiewicz inequality. The proposed algorithms are tested on nonnegative matrix and tensor factorization, as well as matrix and tensor recovery from incomplete observations. The tests include synthetic data and hyperspectral data, as well as image sets from the CBCL and ORL databases. Compared to the existing state-of-the-art algorithms, the proposed algorithms demonstrate superior performance in ...

1,153 citations

Journal ArticleDOI
TL;DR: Several algorithms achieving logarithmic regret are proposed, which besides being more general are also much more efficient to implement, and give rise to an efficient algorithm based on the Newton method for optimization, a new tool in the field.
Abstract: In an online convex optimization problem a decision-maker makes a sequence of decisions, i.e., chooses a sequence of points in Euclidean space, from a fixed feasible set. After each point is chosen, it encounters a sequence of (possibly unrelated) convex cost functions. Zinkevich (ICML 2003) introduced this framework, which models many natural repeated decision-making problems and generalizes many existing problems such as Prediction from Expert Advice and Cover's Universal Portfolios. Zinkevich showed that a simple online gradient descent algorithm achieves additive regret $O(\sqrt{T})$ , for an arbitrary sequence of T convex cost functions (of bounded gradients), with respect to the best single decision in hindsight. In this paper, we give algorithms that achieve regret O(log?(T)) for an arbitrary sequence of strictly convex functions (with bounded first and second derivatives). This mirrors what has been done for the special cases of prediction from expert advice by Kivinen and Warmuth (EuroCOLT 1999), and Universal Portfolios by Cover (Math. Finance 1:1---19, 1991). We propose several algorithms achieving logarithmic regret, which besides being more general are also much more efficient to implement. The main new ideas give rise to an efficient algorithm based on the Newton method for optimization, a new tool in the field. Our analysis shows a surprising connection between the natural follow-the-leader approach and the Newton method. We also analyze other algorithms, which tie together several different previous approaches including follow-the-leader, exponential weighting, Cover's algorithm and gradient descent.

1,124 citations

Journal ArticleDOI
TL;DR: A large deviation-type approximation, referred to as “Bernstein approximation,” of the chance constrained problem is built that is convex and efficiently solvable and extended to the case of ambiguous chance constrained problems, where the random perturbations are independent with the collection of distributions known to belong to a given convex compact set.
Abstract: We consider a chance constrained problem, where one seeks to minimize a convex objective over solutions satisfying, with a given close to one probability, a system of randomly perturbed convex constraints. This problem may happen to be computationally intractable; our goal is to build its computationally tractable approximation, i.e., an efficiently solvable deterministic optimization program with the feasible set contained in the chance constrained problem. We construct a general class of such convex conservative approximations of the corresponding chance constrained problem. Moreover, under the assumptions that the constraints are affine in the perturbations and the entries in the perturbation vector are independent-of-each-other random variables, we build a large deviation-type approximation, referred to as “Bernstein approximation,” of the chance constrained problem. This approximation is convex and efficiently solvable. We propose a simulation-based scheme for bounding the optimal value in the chance constrained problem and report numerical experiments aimed at comparing the Bernstein and well-known scenario approximation approaches. Finally, we extend our construction to the case of ambiguous chance constrained problems, where the random perturbations are independent with the collection of distributions known to belong to a given convex compact set rather than to be known exactly, while the chance constraint should be satisfied for every distribution given by this set.

1,099 citations


Network Information
Related Topics (5)
Optimization problem
96.4K papers, 2.1M citations
92% related
Scheduling (computing)
78.6K papers, 1.3M citations
83% related
Robustness (computer science)
94.7K papers, 1.6M citations
83% related
Linear system
59.5K papers, 1.4M citations
82% related
Nonlinear system
208.1K papers, 4M citations
81% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202311
202236
2021271
2020297
2019268
2018247