Optimization Over Second Order Cones With Extensions To Symmetric Cone Programming PDF Download

Are you looking for read ebook online? Search for your book and save it on your Kindle device, PC, phones or tablets. Download Optimization Over Second Order Cones With Extensions To Symmetric Cone Programming PDF full book. Access full book title Optimization Over Second Order Cones With Extensions To Symmetric Cone Programming.

Reduction Methods in Semidefinite and Conic Optimization

Reduction Methods in Semidefinite and Conic Optimization
Author: Frank Noble Permenter
Publisher:
Total Pages: 266
Release: 2017
Genre:
ISBN:

Download Reduction Methods in Semidefinite and Conic Optimization Book in PDF, ePub and Kindle

Conic optimization, or cone programming, is a subfield of convex optimization that includes linear, second-order cone, and semidefinite programming as special cases. While conic optimization problems arise in a diverse set of fields (including machine learning, robotics, and finance), efficiently solving them remains an active area of research. Developing methods that detect and exploit useful structure-such as symmetry, sparsity, or degeneracy-is one research topic. Such methods include facial and symmetry reduction, which have been successful in several applications, often reducing solve time by orders of magnitude. Nevertheless, theoretical and practical barriers preclude their general purpose use: to our knowledge, no solver uses facial or symmetry reduction as an automatic preprocessing step. This thesis addresses some of these barriers in three parts: the first develops more practical facial reduction techniques, the second proposes a more powerful and computationally efficient generalization of symmetry reduction (which we call Jordan reduction), and the third specializes techniques to convex relaxations of polynomial optimization problems. Throughout, we place emphasis on semidefinite programs and, more generally, optimization problems over symmetric cones. We also present computational results.


Convex Optimization Via Domain-driven Barriers and Primal-dual Interior-point Methods

Convex Optimization Via Domain-driven Barriers and Primal-dual Interior-point Methods
Author: Mehdi Karimi
Publisher:
Total Pages: 139
Release: 2017
Genre: Convex functions
ISBN:

Download Convex Optimization Via Domain-driven Barriers and Primal-dual Interior-point Methods Book in PDF, ePub and Kindle

This thesis studies the theory and implementation of infeasible-start primal-dual interior-point methods for convex optimization problems. Convex optimization has applications in many fields of engineering and science such as data analysis, control theory, signal processing, relaxation and randomization, and robust optimization. In addition to strong and elegant theories, the potential for creating efficient and robust software has made convex optimization very popular. Primal-dual algorithms have yielded efficient solvers for convex optimization problems in conic form over symmetric cones (linear-programming (LP), second-order cone programming (SOCP), and semidefinite programing (SDP)). However, many other highly demanded convex optimization problems lack comparable solvers. To close this gap, we have introduced a general optimization setup, called \emph{Domain-Driven}, that covers many interesting classes of optimization. Domain-Driven means our techniques are directly applied to the given ``good" formulation without a forced reformulation in a conic form. Moreover, this approach also naturally handles the cone constraints and hence the conic form. A problem is in the Domain-Driven setup if it can be formulated as minimizing a linear function over a convex set, where the convex set is equipped with an efficient self-concordant barrier with an easy-to-evaluate Legendre-Fenchel conjugate. We show how general this setup is by providing several interesting classes of examples. LP, SOCP, and SDP are covered by the Domain-Driven setup. More generally, consider all convex cones with the property that both the cone and its dual admit efficiently computable self-concordant barriers. Then, our Domain-Driven setup can handle any conic optimization problem formulated using direct sums of these cones and their duals. Then, we show how to construct interesting convex sets as the direct sum of the epigraphs of univariate convex functions. This construction, as a special case, contains problems such as geometric programming, $p$-norm optimization, and entropy programming, the solutions of which are in great demand in engineering and science. Another interesting class of convex sets that (optimization over it) is contained in the Domain-Driven setup is the generalized epigraph of a matrix norm. This, as a special case, allows us to minimize the nuclear norm over a linear subspace that has applications in machine learning and big data. Domain-Driven setup contains the combination of all the above problems; for example, we can have a problem with LP and SDP constraints, combined with ones defined by univariate convex functions or the epigraph of a matrix norm. We review the literature on infeasible-start algorithms and discuss the pros and cons of different methods to show where our algorithms stand among them. This thesis contains a chapter about several properties for self-concordant functions. Since we are dealing with general convex sets, many of these properties are used frequently in the design and analysis of our algorithms. We introduce a notion of duality gap for the Domain-Driven setup that reduces to the conventional duality gap if the problem is a conic optimization problem, and prove some general results. Then, to solve our problems, we construct infeasible-start primal-dual central paths. A critical part in achieving the current best iteration complexity bounds is designing algorithms that follow the path efficiently. The algorithms we design are predictor-corrector algorithms. Determining the status of a general convex optimization problem (as being unbounded, infeasible, having optimal solutions, etc.) is much more complicated than that of LP. We classify the possible status (seven possibilities) for our problem as: solvable, strictly primal-dual feasible, strictly and strongly primal infeasible, strictly and strongly primal unbounded, and ill-conditioned. We discuss the certificates our algorithms return (heavily relying on duality) for each of these cases and analyze the number of iterations required to return such certificates. For infeasibility and unboundedness, we define a weak and a strict detector. We prove that our algorithms return these certificates (solve the problem) in polynomial time, with the current best theoretical complexity bounds. The complexity results are new for the infeasible-start models used. The different patterns that can be detected by our algorithms and the iteration complexity bounds for them are comparable to the current best results available for infeasible-start conic optimization, which to the best of our knowledge is the work of Nesterov-Todd-Ye (1999). In the applications, computation, and software front, based on our algorithms, we created a Matlab-based code, called DDS, that solves a large class of problems including LP, SOCP, SDP, quadratically-constrained quadratic programming (QCQP), geometric programming, entropy programming, and more can be added. Even though the code is not finalized, this chapter shows a glimpse of possibilities. The generality of the code lets us solve problems that CVX (a modeling system for convex optimization) does not even recognize as convex. The DDS code accepts constraints representing the epigraph of a matrix norm, which, as we mentioned, covers minimizing the nuclear norm over a linear subspace. For acceptable classes of convex optimization problems, we explain the format of the input. We give the formula for computing the gradient and Hessian of the corresponding self-concordant barriers and their Legendre-Fenchel conjugates, and discuss the methods we use to compute them efficiently and robustly. We present several numerical results of applying the DDS code to our constructed examples and also problems from well-known libraries such as the DIMACS library of mixed semidefinite-quadratic-linear programs. We also discuss different numerical challenges and our approaches for removing them.


Lectures on Modern Convex Optimization

Lectures on Modern Convex Optimization
Author: Aharon Ben-Tal
Publisher: SIAM
Total Pages: 500
Release: 2001-01-01
Genre: Technology & Engineering
ISBN: 0898714915

Download Lectures on Modern Convex Optimization Book in PDF, ePub and Kindle

Here is a book devoted to well-structured and thus efficiently solvable convex optimization problems, with emphasis on conic quadratic and semidefinite programming. The authors present the basic theory underlying these problems as well as their numerous applications in engineering, including synthesis of filters, Lyapunov stability analysis, and structural design. The authors also discuss the complexity issues and provide an overview of the basic theory of state-of-the-art polynomial time interior point methods for linear, conic quadratic, and semidefinite programming. The book's focus on well-structured convex problems in conic form allows for unified theoretical and algorithmical treatment of a wide spectrum of important optimization problems arising in applications.


Convex Optimization

Convex Optimization
Author: Stephen P. Boyd
Publisher: Cambridge University Press
Total Pages: 744
Release: 2004-03-08
Genre: Business & Economics
ISBN: 9780521833783

Download Convex Optimization Book in PDF, ePub and Kindle

Convex optimization problems arise frequently in many different fields. This book provides a comprehensive introduction to the subject, and shows in detail how such problems can be solved numerically with great efficiency. The book begins with the basic elements of convex sets and functions, and then describes various classes of convex optimization problems. Duality and approximation techniques are then covered, as are statistical estimation techniques. Various geometrical problems are then presented, and there is detailed discussion of unconstrained and constrained minimization problems, and interior-point methods. The focus of the book is on recognizing convex optimization problems and then finding the most appropriate technique for solving them. It contains many worked examples and homework exercises and will appeal to students, researchers and practitioners in fields such as engineering, computer science, mathematics, statistics, finance and economics.


Handbook on Semidefinite, Conic and Polynomial Optimization

Handbook on Semidefinite, Conic and Polynomial Optimization
Author: Miguel F. Anjos
Publisher: Springer Science & Business Media
Total Pages: 955
Release: 2011-11-19
Genre: Business & Economics
ISBN: 1461407699

Download Handbook on Semidefinite, Conic and Polynomial Optimization Book in PDF, ePub and Kindle

Semidefinite and conic optimization is a major and thriving research area within the optimization community. Although semidefinite optimization has been studied (under different names) since at least the 1940s, its importance grew immensely during the 1990s after polynomial-time interior-point methods for linear optimization were extended to solve semidefinite optimization problems. Since the beginning of the 21st century, not only has research into semidefinite and conic optimization continued unabated, but also a fruitful interaction has developed with algebraic geometry through the close connections between semidefinite matrices and polynomial optimization. This has brought about important new results and led to an even higher level of research activity. This Handbook on Semidefinite, Conic and Polynomial Optimization provides the reader with a snapshot of the state-of-the-art in the growing and mutually enriching areas of semidefinite optimization, conic optimization, and polynomial optimization. It contains a compendium of the recent research activity that has taken place in these thrilling areas, and will appeal to doctoral students, young graduates, and experienced researchers alike. The Handbook’s thirty-one chapters are organized into four parts: Theory, covering significant theoretical developments as well as the interactions between conic optimization and polynomial optimization; Algorithms, documenting the directions of current algorithmic development; Software, providing an overview of the state-of-the-art; Applications, dealing with the application areas where semidefinite and conic optimization has made a significant impact in recent years.


Convex Optimization & Euclidean Distance Geometry

Convex Optimization & Euclidean Distance Geometry
Author: Jon Dattorro
Publisher: Meboo Publishing USA
Total Pages: 776
Release: 2005
Genre: Mathematics
ISBN: 0976401304

Download Convex Optimization & Euclidean Distance Geometry Book in PDF, ePub and Kindle

The study of Euclidean distance matrices (EDMs) fundamentally asks what can be known geometrically given onlydistance information between points in Euclidean space. Each point may represent simply locationor, abstractly, any entity expressible as a vector in finite-dimensional Euclidean space.The answer to the question posed is that very much can be known about the points;the mathematics of this combined study of geometry and optimization is rich and deep.Throughout we cite beacons of historical accomplishment.The application of EDMs has already proven invaluable in discerning biological molecular conformation.The emerging practice of localization in wireless sensor networks, the global positioning system (GPS), and distance-based pattern recognitionwill certainly simplify and benefit from this theory.We study the pervasive convex Euclidean bodies and their various representations.In particular, we make convex polyhedra, cones, and dual cones more visceral through illustration, andwe study the geometric relation of polyhedral cones to nonorthogonal bases biorthogonal expansion.We explain conversion between halfspace- and vertex-descriptions of convex cones,we provide formulae for determining dual cones,and we show how classic alternative systems of linear inequalities or linear matrix inequalities and optimality conditions can be explained by generalized inequalities in terms of convex cones and their duals.The conic analogue to linear independence, called conic independence, is introducedas a new tool in the study of classical cone theory; the logical next step in the progression:linear, affine, conic.Any convex optimization problem has geometric interpretation.This is a powerful attraction: the ability to visualize geometry of an optimization problem.We provide tools to make visualization easier.The concept of faces, extreme points, and extreme directions of convex Euclidean bodiesis explained here, crucial to understanding convex optimization.The convex cone of positive semidefinite matrices, in particular, is studied in depth.We mathematically interpret, for example,its inverse image under affine transformation, and we explainhow higher-rank subsets of its boundary united with its interior are convex.The Chapter on "Geometry of convex functions",observes analogies between convex sets and functions:The set of all vector-valued convex functions is a closed convex cone.Included among the examples in this chapter, we show how the real affinefunction relates to convex functions as the hyperplane relates to convex sets.Here, also, pertinent results formultidimensional convex functions are presented that are largely ignored in the literature;tricks and tips for determining their convexityand discerning their geometry, particularly with regard to matrix calculus which remains largely unsystematizedwhen compared with the traditional practice of ordinary calculus.Consequently, we collect some results of matrix differentiation in the appendices.The Euclidean distance matrix (EDM) is studied,its properties and relationship to both positive semidefinite and Gram matrices.We relate the EDM to the four classical axioms of the Euclidean metric;thereby, observing the existence of an infinity of axioms of the Euclidean metric beyondthe triangle inequality. We proceed byderiving the fifth Euclidean axiom and then explain why furthering this endeavoris inefficient because the ensuing criteria (while describing polyhedra)grow linearly in complexity and number.Some geometrical problems solvable via EDMs,EDM problems posed as convex optimization, and methods of solution arepresented;\eg, we generate a recognizable isotonic map of the United States usingonly comparative distance information (no distance information, only distance inequalities).We offer a new proof of the classic Schoenberg criterion, that determines whether a candidate matrix is an EDM. Our proofrelies on fundamental geometry; assuming, any EDM must correspond to a list of points contained in some polyhedron(possibly at its vertices) and vice versa.It is not widely known that the Schoenberg criterion implies nonnegativity of the EDM entries; proved here.We characterize the eigenvalues of an EDM matrix and then devisea polyhedral cone required for determining membership of a candidate matrix(in Cayley-Menger form) to the convex cone of Euclidean distance matrices (EDM cone); \ie,a candidate is an EDM if and only if its eigenspectrum belongs to a spectral cone for EDM^N.We will see spectral cones are not unique.In the chapter "EDM cone", we explain the geometric relationship betweenthe EDM cone, two positive semidefinite cones, and the elliptope.We illustrate geometric requirements, in particular, for projection of a candidate matrixon a positive semidefinite cone that establish its membership to the EDM cone. The faces of the EDM cone are described,but still open is the question whether all its faces are exposed as they are for the positive semidefinite cone.The classic Schoenberg criterion, relating EDM and positive semidefinite cones, isrevealed to be a discretized membership relation (a generalized inequality, a new Farkas''''''''-like lemma)between the EDM cone and its ordinary dual. A matrix criterion for membership to the dual EDM cone is derived thatis simpler than the Schoenberg criterion.We derive a new concise expression for the EDM cone and its dual involvingtwo subspaces and a positive semidefinite cone."Semidefinite programming" is reviewedwith particular attention to optimality conditionsof prototypical primal and dual conic programs,their interplay, and the perturbation method of rank reduction of optimal solutions(extant but not well-known).We show how to solve a ubiquitous platonic combinatorial optimization problem from linear algebra(the optimal Boolean solution x to Ax=b)via semidefinite program relaxation.A three-dimensional polyhedral analogue for the positive semidefinite cone of 3X3 symmetricmatrices is introduced; a tool for visualizing in 6 dimensions.In "EDM proximity"we explore methods of solution to a few fundamental and prevalentEuclidean distance matrix proximity problems; the problem of finding that Euclidean distance matrix closestto a given matrix in the Euclidean sense.We pay particular attention to the problem when compounded with rank minimization.We offer a new geometrical proof of a famous result discovered by Eckart \& Young in 1936 regarding Euclideanprojection of a point on a subset of the positive semidefinite cone comprising all positive semidefinite matriceshaving rank not exceeding a prescribed limit rho.We explain how this problem is transformed to a convex optimization for any rank rho.


Optimization Over Symmetric Cones Under Uncertainty

Optimization Over Symmetric Cones Under Uncertainty
Author: Baha' M. Alzalg
Publisher:
Total Pages:
Release: 2011
Genre:
ISBN: 9781267198754

Download Optimization Over Symmetric Cones Under Uncertainty Book in PDF, ePub and Kindle

We introduce and study two-stage stochastic symmetric programs (SSPs) with recourse to handle uncertainty in data defining (deterministic) symmetric programs in which a linear function is minimized over the intersection of an affine set and a symmetric cone. We present a logarithmic barrier decomposition-based interior point algorithm for solving these problems and prove its polynomial complexity. Our convergence analysis proceeds by showing that the log barrier associated with the recourse function of SSPs behaves as a strongly self-concordant barrier and forms a self-concordant family on the first stage solutions. Since our analysis applies to all symmetric cones, this algorithm extends Zhao's results [48] for two-stage stochastic linear programs, and Mehrotra and Ozevin's results [25] for two-stage stochastic semidefinite programs (SSDPs). We also present another class of polynomial-time decomposition algorithms for SSPs based on the volumetric barrier. While this extends the work of Ariyawansa and Zhu [10] for SSDPs, our analysis is based on utilizing the advantage of the special algebraic structure associated with the symmetric cone not utilized in [10]. As a consequence, we are able to significantly simplify the proofs of central results. We then describe four applications leading to the SSP problem where, in particular, the underlying symmetric cones are second-order cones and rotated quadratic cones.


Combinatorial and Algorithmic Mathematics

Combinatorial and Algorithmic Mathematics
Author: Baha Alzalg
Publisher: John Wiley & Sons, Inc.
Total Pages: 551
Release: 2024-08-01
Genre: Mathematics
ISBN:

Download Combinatorial and Algorithmic Mathematics Book in PDF, ePub and Kindle

This book provides an insightful and modern treatment of combinatorial and algorithmic mathematics, with an elegant transition from mathematical foundations to optimization. It is designed for mathematics, computer science, and engineering students. The book is crowned with modern optimization methodologies. Without the optimization part, the book can be used as a textbook in a one- or two-term undergraduate course in combinatorial and algorithmic mathematics. The optimization part can be used in a one-term high-level undergraduate course, or a low- to medium-level graduate course. The book spans xv+527 pages across 12 chapters, featuring 391 LaTeX pictures, 108 tables, and 218 illustrative examples. There are also 159 nontrivial exercises included at the end of the chapters, with complete solutions included at the end of the book. Complexity progressively grows, building upon previously introduced concepts. The book includes traditional topics as well as cutting-edge topics in modern optimization.