Science topic

Convex Optimization - Science topic

Explore the latest questions and answers in Convex Optimization, and find Convex Optimization experts.
Questions related to Convex Optimization
  • asked a question related to Convex Optimization
Question
2 answers
Schauder Fixed Point conjecture deals with the existence of fixed points for certain types of operators on Banach spaces. It suggests that every non-expansive mapping of a non-empty convex, weakly compact subset of a Banach space into itself has a fixed point. The status of this conjecture may depend on the specific assumptions and settings.
Relevant answer
Answer
A search with keywords "weak fixed point property" (which is the official name of the property you are interested in) and with "weak normal structure" (which is a widely used sufficient condition for this property) may give you a lot of information on the subject.
  • asked a question related to Convex Optimization
Question
4 answers
Please explain briefly.
Relevant answer
Answer
In general, integer problems are not convex. First of all, due to nonconvexity of the feasible set. However, there are linear integer programming problems possessing properties similar to continuous linear problems. For example, classical transportation, assignment and many network flow problems. The reason is that all vertices of the feasible set have integer components. See more, e.g., in L.Wolsey, Integer Programming.
  • asked a question related to Convex Optimization
Question
3 answers
The steepest descent method proposed by Fliege et al. motivates the research on descent methods for multiobjective optimization, which has received increasing attention in recent years. In the context of convex cases, all the Pareto critical point can be obtained by weighted scalarization method with cheap iterates. Therefore, is it necessary to develop descent methods for convex multiobjective optimization problems?
Relevant answer
Answer
The development of descent methods for convex multiobjective optimization problems may still be useful, even in cases where the Pareto critical points can be obtained using weighted scalarization.
One reason for this is that descent methods can often be implemented more efficiently and with lower computational complexity than weighted scalarization. This can make them more practical to use in cases where the optimization problem is large or the available computational resources are limited.
Additionally, descent methods may be more flexible than weighted scalarization, as they can be applied to a broader range of optimization problems and can be adapted to different scenarios and objectives. This can make them more versatile and useful in a wider range of applications.
Overall, while weighted scalarization can be a useful tool for solving convex multiobjective optimization problems, the development of descent methods for these problems is still an active area of research and can provide valuable insights and solutions for addressing these types of optimization challenges.
  • asked a question related to Convex Optimization
Question
5 answers
linear least squares problem::
fit f(t) = sin(t) on the interval [0, π] with polynomials of degree n with n = 1, 2, 3, · · · , 10. Here, we use equally spaced nodes.
Solve the normal equation with a backslash in Matlab. Save the norm of the residual and the condition numbers of AT A??
Could anybody please tell me how can find x, y, and A in that case ??
Relevant answer
Answer
The question seems incorrect. We needn't find x and y.
We have {(xi, yi = sin(xi))}, i=1,...,N and seek (a_n, ..., a_0) for a polynomial function
f(x) = a_n * x^n + ... + a_0.
Least Square minimization gives
A=
( Sum(xi^n) Sum(xi^(n-1)) ... Sum(xi^0);
Sum(xi^(n+1)) Sum(xi^(n-2)) ... Sum(xi^1);
.............................................
Sum(xi^(2*n)) Sum(xi^(2*n-1)) ... Sum(xi^n);
);
b = (Sum(yi); Sum(xi*yi);); ... Sum(xi^n*yi) );
At last,
(a_n a_(n-1) ... a_0) = A^(-1)*b
  • asked a question related to Convex Optimization
Question
2 answers
Dear Scholars,
I would like to solve a Fluid Mechanic optimization problem that requires the implementation of an optimization algorithm together with Artificial Neural network. I had some questions about Convex optimization algorithm and I would appreciate it if you could give some advice to me.
My question is about possibility of implementing of the Convex optimization together with Artificial Neural Network to find a unique solution for a multi-objective optimization problem. The optimization problem that I am trying to code is explained as the following equations. The objective function utilized in the optimization problem is defined in the following equation:
📷
Where OF is the objective function, wi are the weights assigned to each cost function, Oi is the ith cost function defined as the relative difference between the experimental and the predicted evaporation metrics for fuel droplet (denoted by superscript of exp and mdl, respectively), k is the number of cost functions (k = 4, equal to the number of evaporation metrics), c [c1, c2, and c3] is the mass fraction vector defining the blend of three components, subjected to the following constrains:
📷
Due to high CPU time required for modeling and calculating the objective functions (OF), an ANN was trained based on some tabulated data from modeling of fuel droplet evaporation and used for calculating the OF through optimization iteration.
In the same manner, the wi values are subjected to optimization during the minimization of OF, with the following constraints:
📷
It worth mentioning that, I have solved this problem by employing Genetic Algorithm together with ANN. Although, the iterative process for solving the problem converged to acceptable solutions. But, the algorithm did not return a unique solution.
Regarding that, I would ask you about possibility of using Convex optimization algorithm together with ANN for solving the aforementioned problem to achieve a unique solution. In case of such feasibility, I would appreciate it if you could mention some relevant publications.
Relevant answer
Answer
Switching your optimization algorithm will probably not give you the unique solution that you are looking for. In general the loss function of a neural network is not convex with respect to the parameters. This means that you will have different local minima or saddle points in your loss function. A convex optimization algorithm will converge to one of these points depending on its starting point. Finding the global minimum is a very hard problem and we don't know how to find it, or how to know if we have found it. This is a well known problem of neural networks. Of course, if you want to get the same solution every time that you run your algorithm you can simply set the initial parameters to a fixed value, so the algorithm always starts at the same place. If you do that, then algorithms like gradient descent will always stop at the same point...
  • asked a question related to Convex Optimization
Question
8 answers
For example :
structures & aerodynamics disciplines.Objective functions: weight, wing displacement and aerodynamic coefficient.
Relevant answer
Answer
I suggest you take a look on: https://www.mdpi.com/2226-4310/9/3/153
where aero-structural optimization is carried out using Particle Swarm Optimization.
  • asked a question related to Convex Optimization
Question
11 answers
The question is about finding convexity of any optimisation problem involving equality constraints. I will start with my belief first so if there are already some issues there, they can be addressed right away.
Belief: A problem is convex iff
1. Functions defining inequality constraints i.e. g_i(x) <=0 are all convex. The easiest way to check this is by finding hessian of all g_i(x) and it should be positive semi-definite (all eigen values must be non-negative).
2. Functions defining equality constraints i.e. h_i(x) = 0 must be all linear.
3. The objective function must be convex "over the feasible set".
Now the issue I am facing is with the 3rd one. I believe that the function does not necessarily need to be convex over R but only over the feasible set defined by g_i(x)<=0 and h_i(x)=0. Now, if we would just have inequality constraints, it would be easy to determine whether objective function is convex or not. We can just find it's hessian and see whether the hessian stays positive semi-definite for all values within the feasible set (If it would be a quadratic function, the hessian would be constant and the values of X will not matter anyway). BUT it is a bit difficult to do this exercise with the equality constraints like how do we test the hessian for values of X which satisfy these equalities.
I will highlight this issue with a simple example:
Let's say we have a problem with objective function
f(X) = -x*y with X = [x;y]
and equality constraint
h_(X)=x+y =0
Now the equality constraint is linear so the feasible set is convex. f(X) however seems to be non-convex at first glance as if we find hessian(f,X), it comes out to be [0,-1;-1,0] which has eigen values as -1 and 1. Thus, the hessian is not positive semi-definite and f(X) is not convex (at least not over R).
On the other hand, if we would choose to eliminate y by using h_(X) i.e. y = -x, we get an f(x) = x^2 which is totally convex over R1. Does this mean that f(X) is convex over the feasible set defined by x+y=0??
If yes, then how should we check the convexity of the objective function when equality constraints are involved? We clearly see that just checking the hessian of f(X) may not be enough!
Relevant answer
Answer
You're right: a local minimum is characterized by the (necessary) KKT conditions and (sufficient) second-order conditions: https://en.wikipedia.org/wiki/Karush%E2%80%93Kuhn%E2%80%93Tucker_conditions#Second-order_sufficient_conditions
They require that the Hessian be SPD on the nullspace of the Jacobian of the active constraints, that is exactly what happens in your example. Since your constraint is an equality and is linear, the Jacobian of the active constraints is the vector (1, 1) and its nullspace is the feasible set itself.
So your conclusion is correct: the Hessian should be SPD on the set {(x, y) | x+y = 0}.
In practice, this condition is not super easy to check, so practitioners often make the Hessian positive definite. Period.
  • asked a question related to Convex Optimization
Question
5 answers
I am interested in global convergence and application of that algorithm to different area.
Any help is highly appreciated
Relevant answer
Answer
Interior point methods (https://en.wikipedia.org/wiki/Interior-point_method) are generally superior to active-set methods (https://en.wikipedia.org/wiki/Active-set_method) for solving large-scale problems.
You usually end up solving a convexified equality-constrained problem, that is a system of nonlinear equations. That's the most favorable case.
Among active-set methods, SLQP (https://en.wikipedia.org/wiki/Sequential_linear-quadratic_programming) is of particular interest. It also forms an equality-constrained quadratic problem (the favorable case) after it approximated the set of active inequality constraints by solving a linear problem.
  • asked a question related to Convex Optimization
Question
7 answers
Hello, I have a basic question,
I know that the convex optimization problem is the one in which its objective and inequality functions are convex and its equality function is affine. My question is which set the objective and inequality functions have to be convex on? Is it their entire domain? or just the feasible set of the problem?
For example: let the problem be minimizing f0(x),
s.t.:
fi(x)<=0. for i = 1, 2, ..., m.
and f1(x), for example, is not a convex function on its entire domain, but it is convex on the feasible set of the problem. Is this problem is meant to be convex?
Relevant answer
Answer
Most results and algorithms of convex optimization would not hold if the functions involved are only convex on the feasible set (provided the latter is convex - as Mounir Haddou pointed out). For example, the Lagrangian will not be convex spoiling most of the duality theory; if we are talking about nonsmooth problems, subgradiens may not exist, etc etc. One can of course rectify this by defining the functions involved to be +infinity outside of the (convex, closed) feasible set thereby restoring their convexity, but this will ruin any smoothness your original problem may have had.
  • asked a question related to Convex Optimization
Question
3 answers
I am trying to minimize the following objective function:
(A-f(B)).^2 + d*TV(B)
A: known 2D image
B: unknown 2D image
f: a non-injective function (e.g. sine function)
d: constant
TV: total variation: sum of the image gradient magnitudes
I am not an expert on these sorts of problems and looking for some hints to start from. Thank you.
Relevant answer
Answer
Om Prakash Yadav Thank you, will take a look.
  • asked a question related to Convex Optimization
Question
10 answers
Hi,
I have a little knowledge about optimization and I have a simple question. If my convex optimization problem has an inequality constraint that ensures that the optimization variable can't be less than zero and greater than a maximum value and I have an iterative algorithm to solve this problem. My question is, can I handle this constraint with programming by forcing the value of the variable to not be less than zero and larger than the maximum value in each iteration (with the min() and max() functions for example)? or I must handle this constraint by two Lagrangian multipliers?
Relevant answer
Answer
Look up "gradient projection method" - which is very simple to implement if you have a feasible set bounded by linear constraints. Moreover, when doing line searches you have a very easy way to control that your iteration points are feasible.
  • asked a question related to Convex Optimization
Question
11 answers
I need to implement a load scheduling algorithm that involves solving an online optimisation problem from a research paper for my Real time systems course.
This convex optimisation problem is setup through Model predictive control or receding horizon control. This problem involves under 100 decision variables. I am just googling about solvers and it would be great if you can give me more accurate suggestions about specific numerical solvers available in python for solving these kinds of cost functions (quadratic + one norm + infinity norm) and constraints (both equality and inequality). I think its convex (single minimum) because the author says so.
*If you are an expert you dont have to read the contents below just click and look at the first attachment and suggest to me a solver or links to a solver in python, Thanks :)*
Please see the cost function and constraints in the attachment.
**Brief description of the problem:**
We want to schedule tasks (i.e. power allocation) that maximises power allocation from renewable sources, W to Electric vehicle (EV) loads, E that have some timing deadlines, di (due to running out of charge). We should minimize the use of grid energy sources, G and only use it in the worst case when EV load demand cannot be satisfied using renewable sources.
In the objective function the matrix W represents renewable energy/power allocation for task i at time step k. we look into N time steps into the future from the current time step, t.
Say we have a renewable energy forecast chart that predicts the maximum output power at future times like in renewable energy generation forecast graph attached. Matrix G is grid generated power/energy also indexed similarly to W and is available at all times. Total power/energy supplied to load E is W + G. phi (flexibility factor) is the term that implicitly brings W into the cost fucntion. we try to keep phi large so that we can always defer load scheduling to future times if renewable power/energy is not sufficiently available for the time being and maybe available in the future. Note: the size/dimension (MXN) of matrix W and G is decreasing as we approach end of active tasks list in the time horizon N. As tasks in active tasks get completed and as we approach the last deadline in active task list M and N are decreasing.
In constraint 7, we put a cap on the maximum renewable energy available for all the M tasks up to N time steps using the renewable energy generation forcast graph attached.
constraint 8 is a conservation equation W and G power/energy supplied equals load energy required, E for the M tasks.
In constriant 9 mi is maximum power that can be supplied to the load due to loads physical limitation.
In contraint 10 and 11, di is the deadline of task i, delta t is time step for within which the optimisation must be solved. ei(k) is the remanining energy required for the task to complete.
From this description please suggest solvers or links to solvers that can be used in python.
Thank you!!
  • asked a question related to Convex Optimization
Question
13 answers
Mathematical programming is the best optimization tool with many years of strong theoretical background. Also, it is demonstrated that it can solve complex optimization problems on the scale of one million design variables, efficiently. Also, the methods are so reliable! Besides, there is mathematical proof for the existence of the solution and the globality of the optimum.
However, in some cases in which there are discontinuities in the objective function, there would be some problems due to non-differentiable problem. Some methods such as sub-gradients are proposed to solve such problems. However, I cannot find many papers in the state-of-the-art of engineering optimization of discontinuous optimization using mathematical programming. Engineers mostly use metaheuristics for such cases.
Can all problems with discontinuities be solved with mathematical programming? Is it easy to implement sub-gradients for large scale industrial problems? Do they work in non-convex problems?
A simple simple example of such a function is attached here.
Relevant answer
Answer
It may be useful to make a distinction between 'global' optimisation (where one is given an explicit, non-convex objective function) and 'black-box' optimisation (where the function is not given explicitly). The methods for them are very different.
  • asked a question related to Convex Optimization
Question
4 answers
What are some of the well-written good references that discusses why finding the penalty parameter when solving a nonlinear constrained optimization problem is hard to find from the computational perspective. What are some of the computational methods done to find the parameter as I understand finding such a parameter is problem-dependant.
Any insights also would be very helpful.
Thanks
Relevant answer
Answer
Some penalty methods are exact, in the sense that a finite penalty is enough to identify an optimum. Typically the penalty function then needs to be non-differentiable, which might be a bummer. With differentiable penalty functions, such as the augmented Lagrangian, you typically need the penalty to grow indefinitely, but there are those that can be exact without a growing penalty parameter. It however has a cost, in the sense that some constraint qualification holds - which is hard to check.
Here is a paper about just that:
Exact augmented Lagrangian functions for nonlinear semidefinite programming, by Ellen H. Fukuda & Bruno F. Lourenço
Computational Optimization and Applications volume 71, pages 457–482 (2018)
Hope it helps! :-)
  • asked a question related to Convex Optimization
Question
5 answers
Is there such a thing as convex optimization for large deep learning networks? Is there a way to guarantee convergence? Has anyone looked into this?
Relevant answer
Answer
Dear Brian,
I think for your initial query you got very solid answers. For your last query, I think, YES you apply manifold learning technique by dividing the non-convex surface into a large collection of convex Euclidean surfaces whose union will be equivalent to the whole non-convex Riemannian surface. Such a procedure may be extended to the application of deep learning or iterative learning via Grassmann manifolds or symmetric positive definite matrices. To study this further, I suggest you to read some related work, e.g., https://arxiv.org/pdf/1411.8003.pdf
  • asked a question related to Convex Optimization
Question
8 answers
in the method (HPM) I do not know how to construct the homotopy convex in many papers i find different of type homotopy (is it modification of these method or what ?)
and what is the conditions for construct the correct homotopyconvex
#applied mathematics #Neumerical analysis #Homotopy perturbation method
Relevant answer
  • asked a question related to Convex Optimization
Question
4 answers
Considering the powered descent guidance problem for a spacecraft or launch vehicle - Early publications by Acikmese, et.al. used lossless convexification to solve the landing guidance problem using point mass dynamics.
Literature that dealt with the powered descent guidance problem using 6 DoF dynamics, also by Acikmese, et.al. proposed the use of successive convexification, with the reasons being that lossless convexification could not handle non-convex state and some classes of non-convex control constraints.
I have the following questions with respect to the above
1) Why is it that Lossless convexification cannot handle some classes of non-convex control constraints ? and what are those classes ? (e.g. the glideslope constraint)
2) What is it about lossless convexification that makes it unsuitable to operate on non-convex state constraints ,as opposed to successive convexification ?
2) Are there specific rotational formalisms ,e.g. quaternions, DCMs, Dual quaternions, MRPs, etc. that are inherently convex ? if so, how does one go about showing that they are convex ?
I would be much obliged if someone could answer these questions or atleast provide references that would aid in my understanding of the topic of convexification.
Thank you.
Relevant answer
Answer
Dear Adhithya Babu,
A successive convexification algorithm designed to solve non-convex constrained optimal control problems with global convergence and it can handle nonlinear dynamics and non-convex state and control constraints. However lossless convexification cannot handle nonlinear dynamics. For the third question the answer is affirmative.
For more details and information about your questions I suggest you to see links and attached files on topic.
Best regards
  • asked a question related to Convex Optimization
Question
5 answers
I would be grateful if anyone could tell me how the McCormick error can be reduced systematically. In fact, I would like to know how we can efficiently recognize and obtain a tighter relaxation for bi-linear terms when we use McCormick envelopes.
For instance, consider the simple optimization problem below. The results show a big McCormick error! Its MATLAB code is attached. Min Z = x^2 - x s.t. -1 <= x <= 3 (optimal: x* = 0.5 , Z* = -0.25 ; McCormick: x*=2.6!)
Relevant answer
Answer
Hi Morteza,
I am a new researcher in this field and I am working on similar problems. I hope you will find following paper useful:
Cheers,
Zaid
  • asked a question related to Convex Optimization
Question
4 answers
I have rewritten an MPC problem to a QP formulation.
The QP solvers quadprog and Gurobi, which uses the interior-point algorithm, give me the same objective function value and optimized values x, but GPAD, a first-order solver, gives me the same optimized values x, but an objective function value, which is a factor 10 bigger compared to quadprog and Gurobi. Do you anyone know a possible explanation?
Relevant answer
Answer
Good!
  • asked a question related to Convex Optimization
Question
3 answers
It seems that the quadprog function of MATLAB, the (conventional) interior-point algorithm, is not fully exploiting the sparsity and structure of the sparse QP formulation based on my results.
In Model Predictive Control, the computational complexity should scale linearly with the prediction horizon N. However, results show that the complexity scales quadratically with the prediction horizon N.
What can be possible explanations?
Relevant answer
Answer
Thanks for your answer.
I gave the wrong information. Quadprog is not exploiting the sparsity and structure of the sparse QP formulation at all. So, the computation complexity scales cubically with N.
The barrier algorithm of Gurobi was not fully exploiting the sparsity and structure of the sparse QP formulation. So, the computation complexity scales cubically with N. Do you have some documentation about this? At the Gurobi website, I could not find anything relevant to answer this question.
Could you maybe also react to my last topic about why interior point algorithm of Gurobi and quadprog give same optimized values x and corresponding objective function value , but the first-order solver GPAD gives the same optimized values x, but another objective function value which is a factor 10 bigger?
Your answers are always really helpful.
Regards
  • asked a question related to Convex Optimization
Question
3 answers
I am trying to solve a QP problem.
Does anybody know the differences between the interior-point-convex algorithm of quadprog and the barrier method of Gurobi in terms which kind of matrices can the solvers handle the best? Sparse matrices or dense matrices? And what kind of problems: large problems or small problems? Furthermore, which solver needs more iterations (with less expensive cost per iteration) and which solver needs fewer iterations, but expensive cost/time per iteration.
From results, I think Gurobi gives more iterations and quadprog less, but I do not know why?
And what are the differences with GPAD by means of what is described above?
Relevant answer
Answer
Dear Mohamed,
Quadprog and Gurobi are given me the same objective function value and optimized values x , but GPAD gives me the same optimized values x, but an objective function value, which is a factor 10 bigger compared to quadprog and Gurobi. How is this possible?
  • asked a question related to Convex Optimization
Question
4 answers
According to first order Taylor series approximation, |xHw|2 can be approximated as:
|xHw|2 >= 2*real{(w0)HxxHw}-|xHw0|2
where x and w is colum vector and w is variable vector which is unknown.
How to find a lower approximate f(w) of |xHw|2*|yHw|2- |zHw|2*|bHw|2 satisfies the follwoing condition:
|xHw|2*|yHw|2- |zHw|2*|bHw|2>= f(w)
where w is variable vector which is unknown, vectors x,y,z,b are known.
Relevant answer
Answer
Dear Renaud Di Francesco ,
Thank you for your answer. I want to solve an optimization problem, such as in the Figure.
The constraint is not convex. So I want to approximate the left-hand expression of the inequality by a lower approximation which is convex or linear on w.
  • asked a question related to Convex Optimization
Question
13 answers
Hello everyone,
I am a PhD student with a deep learning background. I am trying to find more robust deep learning models and I just came through defining the problem as a minimax problem as the attacker tries to maximize the loss function and the defender tries to minimize the risk of he attacker. I have just learned that this problem can be solved in three ways : 1- Lower bounds 2- Exact solutions 3- Upper bounds.
The problem is that I do not know how to get deeper in optimization in order to get a unique solution to the problem.
Can you recommend some resources or road-map for optimization, convex optimization, non-convex optimization, and robust optimization?
Relevant answer
Answer
Bi-level Mini-max problems can be solved with Benders (1962) decomposition. The inner maximization problem becomes the "subproblem" and the outer minimization becomes the "master" problem.
If the inner problem has all continuous variables, then you can try to formulate the dual of the inner maximization and get the inner minimzaiton of the dual variables. If the inner subproblem dual is possible, this would reduce your mini-max problem to a single level minimization problem.
  • asked a question related to Convex Optimization
Question
2 answers
Hi all,
Is ADMM (Alternating Direction Method of Multipliers) numerically stable? That is, if there is a small error in a problem parameter, would the error in the obtained solution be small as well?
If it is stable, can you showed me a paper that proves the stability? If it is not, do you know any extension to the algorithm that makes it stable?
Thanks,
Alex
  • asked a question related to Convex Optimization
Question
3 answers
I have found many related optimization references, but there were no method or enlightments for my specific optimization problems.
known: D, W, P_max, h_j, T_k, \sigma_u.
Unknown: p_k. k=1,2,3,...K. (for example, K=3 users)
Here, I intend to find the optimization variables of $\mathbf p = [p_1, p_k, ..., p_K]^{T}$。
Hope for some suggestions or enlightments from convex optimization researchers.
Thanks a lot!
Relevant answer
Answer
This is a bounded variables optimization problem, also the value of the logarithm can be calculated at each k. So you can solve it using WINQSB Solver.
  • asked a question related to Convex Optimization
Question
5 answers
I have a relaxed convex problem (Conic) that can be expressed using two formulations. When solved using CPLEX, I obtained the same global solution. However, one formulation yielded the global solution in much faster time than the other. My question is how to explain these finding scientifically? My guess is that the faster algorithm has a smaller feasible space which allow the solver to find the solution in a faster way but I can't prove this or even know if this is really the case. Any suggestion that can help in comparing the performance?
Relevant answer
Answer
Miguel: in decomposition methods one does far more work than evaluating the objective function, such as solving very, very many simpler sub-problems; hence in order to estimate the time to solve an instance of the problem we cannot simply sum up the number of original objective function evaluations.
  • asked a question related to Convex Optimization
Question
12 answers
Is linear programming one part of convex optimization?
Relevant answer
Answer
Convex optimization involves minimizing a convex objective function (or maximizing a concave objective function) over a convex set of constraints.
Linear programming is a special case of convex optimization where the objective function is linear and the constraints consist of linear equalities and inequalities.
Nonlinear programming concerns optimization where at least one of the objective function and constraints is nonlinear.
(Adapted from Mathematical optimization: Major subfields on Wikipedia.)
Therefore, convex optimization overlaps both linear and nonlinear programming, being a proper superset of the former and a proper subset of the latter. However, note that nonlinear programming, while technically including convex optimization (and excluding linear programming), can be used to refer to situations where the problem is not known to be convex (see Boyd and Vandenberghe, p. 9, below). Hence, it may be more useful in practice to think of a hierarchy: linear - convex - nonlinear. Another useful view is given by the following quote, kindly supplied by littleO: "The great watershed in optimization isn't between linearity and nonlinearity, but convexity and nonconvexity." -- R. Tyrrell Rockafellar, in SIAM Review, 1993
  • asked a question related to Convex Optimization
Question
4 answers
Deterministic Global optimization relies on convex relaxation of the non-convex problems. Certain nonlinearities are duly converted into linear forms underestimators to be solved by efficient MILP solvers (e.g. signomial functions/ bilinear terms).
Most nonlinearityies are approximated to linear functions by piece-wise linearizations. However, I am wondering if this linearizations guarantees that the approximations are understimators of the original nonconvex problem (i.e. for all x in Domf, f(x) >= u(x) where u is the understimator)
because otherwise the understimator may miss the global optimum during the branch and bound process.
Can the solver still converge even if the relaxation is not an understimator?
Relevant answer
Answer
That's right. More precisely, if an inequality is of "<=" type, we use convex under-estimators for the left-hand side. If the inequality is of ">=" type, we use concave over-estimators (or you can just multiply the inequality by -1).
McCormick's original paper (from 1972) explains this very well. So does this paper:
Ryoo, H. S., & Sahinidis, N. V. (1996). A branch-and-reduce approach to global optimization. Journal of global optimization, 8(2), 107-138.
  • asked a question related to Convex Optimization
Question
6 answers
min - x - y
s.t. xy <= 4
0<= x <= 4
0 <= y <= 8
Relevant answer
Answer
The following problem can be transformed into Geometric Programming as follows: Minimize -x-y
xy ≤ 4
0 ≤x ≤ 4
0 ≤ y ≤ 8
Consider the transformation :
X = √(4-x)+√(4+y) and Y = √(4-x)-√(4+y)
This transforms - x - y = XY
where x = 4 - (X+Y)2/4 ,and y = 4 - (X - Y)2/4
The problem transforms into (Geometric Programming)
Minimize f0(X,Y) = XY
subjected to the constraints:
xy ≤ 4 ⇒ (4 - (X+Y)2/4)(4 - (X - Y)2/4) ≤ 4
f1(X,Y) = (1 - (X+Y)2/16)(1 - (X - Y)2/16) ≤ 1.
0 ≤ x ≤ 4 ⇒ 0 ≤ 4 - (X+Y)2/4 ≤ 4 ⇒ f2(X,Y) = 1 - (X+Y)2/16 ≤ 1.
0 ≤ y ≤ 8 ⇒ 0 ≤ 4 - (X - Y)2/4 ≤ 8 ⇒ f3(X,Y) = 1/2 - (X - Y)2/32 ≤ 1.
Best regards
  • asked a question related to Convex Optimization
Question
6 answers
Please give their appropriate cases.
Relevant answer
Answer
we can chose linear or dual dependent to conditional constrants
  • asked a question related to Convex Optimization
Question
13 answers
The constraints include both linear constraints and nonlinear constraints. The essential issue lies in to how to deal with the nonlinear constraints.
It would be better if this algorithm can transform these nonlinear constraints into the equivalent linear ones.
Relevant answer
  • asked a question related to Convex Optimization
Question
4 answers
How can implementation an Adaptive Dictionary Reconstruction for Compressed Sensing of ECG Signals and how can an analysis of the overall power consumption of the proposed ECG compression framework as would be used in WBAN.
Relevant answer
Answer
Thid is in class of machine learning in youtubre
  • asked a question related to Convex Optimization
Question
5 answers
To solve a highly nonlinear function. We can not show analytically this function is quasi-concave.
Relevant answer
Answer
you prove two things:
1) dom(f) is convex
2) all sublevel sets of -f(x) are convex
  • asked a question related to Convex Optimization
Question
5 answers
NMPC- Nonlinear model predictive control
X_dot=f(x,u)
Y=C*x
objective function: min J = (Y-Ys)^2+du^2+u^2
w.r.t u
constraints are :
0<u<10
-0.2<du<0.2
0<Y<8 here Y is a nonlinear constraint and a vector
Relevant answer
Answer
Hi,
Here is a complete workshop on how to implement MPC and MHE in MATALB.
The workshop shows a complete explanation of the implementation with coding examples. the codes are also provided
  • asked a question related to Convex Optimization
Question
13 answers
Dear Friends and colleagues
I have an optimization in which I have a nonlinear term in the following form:
x(t)* a(k)
where, x and a are variables. a is a binery variable and the sets in which each of the variables are defined is not the same. Would you please suggest me a method that I can use to handle this term and transfer my model to a mixed integer linear programming?
Thank you for your suggestions.
Relevant answer
Answer
Olivér Ősz is right.
  • asked a question related to Convex Optimization
Question
29 answers
best (achievable) x-reg for singular solutions of nd incompressible euler equations by Székelyhidi jr.+de lellis and remarks/links with pressureless (regular) sols ?
--on solutions "convex optimized" by l. Székelyhidi jr. and c de lellis (and descendants) of nd (or 2d, 3d if there is a difference for LS-CD frames and if so, then how ?) incompressible euler equations, could anyone tell what is the best x-reg (so better than C ^ s, s about 1/3) that LSjr+CD or their descendants, 1 / have already, 2 / could be expected and what are the obstacles getting (or not) a better x-reg ? (all with the same question in t-reg and mixed (x, t) -regs), this regardless of their first motivations, that is, breaking the uniqueness (more generally realistic up to C ^ s for all s <1) and onsager (related to s=1/3) or the final answer is close to s=1/3 (and then why)?
--Can anyone confirm that the solutions coming from the classical and regular theory (that is to say not LS&CD theory), which are the particular solutions that are  pressureless (see in 2 / 3d, majda diperna pl lions but also in Rn, all n, in my jmpa95 and thesis92-ch3/90 etc, quoted by yudovich) are out of reach by LSjr+CD theory and their singular frameworks (because for LSjr+CD, the pressure is, at first, basically + - equal to cst lul ^ 2 and then p = 0 gives u = 0 or all of this can be (partly) overcome and how, for example, include these sols in LS&CD setting?) = What could be said about these sols regarding the LS&CD theory ?
--regular theory =eg, for a (short) interval of non nul times, +- Du or rotu (:Du-tDu) are in x, C^s (s in ]0,1[), C^o, or L°° or bmo or (eg in 2d:) Du or rotu in Lp(loc), for one p in [1,oo] etc
--LS&CD singular theory at the state (?)= eg , for a (short) interval of non nul times, u itself (and not its gradient Du) is only in C^s, s in ]0,1[ or s<1/3 or s= 1/3 etc and eg nothing (yet?) on (atleast L1(loc)/loc-measures) DERIVATIVES of u, eg Du or rot(u) obtained (and conserved) for non nul times (even if it is supposed at t=0) etc
-- Are there (zones of) junctions or intersections (and where and what) between regular theory and LS&CD singular theory, existing same common sols (even particular) for these 2 theories ?
--eg 2d, or 3daxi, rotu (and/or its moments 0,1,2 and/or the axi structure) could be? (even "half" or "under" or partly in some way) conserved in SDth as in regTh etc ?.
--questions extended of course to all flu mech models already treated (by extensions from IncEE) by SDth. 6/7/18
Relevant answer
Answer
on the comment of P. K. Karmakar = the sols and theory by
Székelyhidi jr+de lellis on, at first, the incEE (and afler, other models of flu mech), date from c2007 (excepted mistake of my part for all), not before, and it is of a hight level of theorical maths (a talk at the bourbaki seminar was decided and made on it) one of the most important results on incEE for the last eg 20 years. the book you said is earlier, ie 2002
  • asked a question related to Convex Optimization
Question
17 answers
what is the best method for constrained optimization ( in terms of the speed of the convergence an accuracy)?
I want to train the parameters of the neural network with some constraints on the output of the network, so I need an constrained optimization method. the number of parameters ( weights and biases of neural network) is about 200 to 1000 parameters.
Relevant answer
Answer
See our method in: N.D. Lagaros, M. Papadrakakis, “Applied soft computing for optimum design of structures”, Structural and Multidisciplinary Optimization, Vol 45, pp. 787-799, 2012. Page: 797. In particular, the simple yet effective, multiple linear segment penalty function is used in this study for handling the constraints. According to this technique if no violation is detected, then no penalty is imposed on the objective function. If any of the constraints is violated, a penalty, relative to the maximum degree of constraints’ violation, is applied to the objective function. On the other hand, the optimization procedure is terminated when the best value of the objective function in the last 20 generations remains unchanged.
  • asked a question related to Convex Optimization
Question
7 answers
I am reading Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers by Boyd (https://web.stanford.edu/~boyd/papers/pdf/admm_distr_stats.pdf), where the authors showed that ADMM converges under 2 assumptions (page 16):
1. the objective function f and g are closed, proper and convex.
2. the unaugmented Lagrangian $L_0$ has a saddle point.
A proof is given in Appendix A in the above PDF (page 106-110).
I am working on a problem which I would like to solve using ADMM, but the objective functions in my problem are not closed (but they are differentiable and convex). In this case, can I still use ADMM and have guaranteed convergence?
I am asking this because from the proof mentioned above, it seems that the assumption 1 is only used to justify the sub-differentiability of $L_\rho$, so I think I can still use ADMM in my problem given that $L_\rho$ is differentiable.
Relevant answer
Answer
Hi Jakub,
Thanks for your answer and you are absolutely right. I forgot the fact that rays (such as [1, \infty)) are closed sets, which makes me think that linear functions are not closed.
Thanks,
Alex
  • asked a question related to Convex Optimization
Question
4 answers
Would you please explain to me what is the technical definition or explanation behind the point that, in a convex optimization program or space, the equality constraint must be affine.
Relevant answer
Answer
A simple explanation:
Let h(x) == constant be the equality constraint in the optimization problem and x is the optimization variable. The above equality constraint can be rewritten in terms of inequality constraints as follows:
h(x) <= constant -- (1)
h(x) >= constant -- (2). If both (1) and (2) are satisfied then it is equivalent to the h(x)== constant.
If h(x) is a convex function then (1) is a convex constraint. However, (2) is not a convex constraint anymore. Similar argument holds when h(x) is a concave function. Thus, in order for both (1) and (2) to be a convex constraint h(x) has to be an affine function.
  • asked a question related to Convex Optimization
Question
8 answers
Hi,
I am working with a VERY large scale LP -- so large that simplex method takes forever to run. I've developed an efficient numerical algorithm to exploit the problem structure to significantly reduce the running time.
The problem is, my application requires basic solutions. For now, I am using crossover, that is, take the optimal primal and dual solutions (numerical) from my algorithm, give them to a simplex solver, and use good old simplex to solve it to get basic solutions. This works well, but it's still a bit too slow, as most of the running time are used in the crossover phase.
So my question is, is there any other algorithm in the literature that can produce basic solution given a numerical solution with high accuracy?
Thanks,
Alex
Relevant answer
Answer
Nice links from Nicolas Dupin. The articles are interesting and straight to the point. Am sure Alex will find them useful.
  • asked a question related to Convex Optimization
Question
6 answers
I am trying to solve SDPs using mosek where IPM is used. For very large SDP problem mosek has some issues. Is there any other solver available which uses different algorithm and can solve large scale SDPs ??
Relevant answer
Answer
I kind of surprised no one mentions Mosek. It is fastest SDP code these days according to
  • asked a question related to Convex Optimization
Question
3 answers
In solving sparse representation problems in image processing generally the L1-minimization techniques gives good result compared to other greedy algorithms. But sparse coding method that solves the minimization based on 'Feature-sign search algorithm' is slow. Does the methods like, iterative shrinkage thresholding algorithm (ISTA) or FISTA (fast ista) which are commonly used in image denoising problems, will be more efficient if it is used to replace the 'Feature-sign search algorithm' in image super-resolution ?
Relevant answer
Answer
DCT-based L1-regularization for robust data interpolation might be an efficient solution. Follow:
In search of efficiency, switching to L2-Regularization could be interesting. To help you choose the best compromise, check out:
Regards
  • asked a question related to Convex Optimization
Question
3 answers
How do we carry iteration of matrix rank approximation and what is the initial value and step size? M is matrix with missing values. We approximate X to M. Its in matrix completion problem. Any answer is appreciable.
Relevant answer
Answer
Thanks to all, for your valuable feedback.
  • asked a question related to Convex Optimization
Question
1 answer
I have studied and understood the Moment-SOS hierarchy proposed by Lasserre where a sequence of semi-definite programs are required to be solved and a rank condition is invoked in order to check if the global solution is found. I was not able to find such conditions for its dual viewpoint ( also known as the Putinar's Positivstellensatz). Alternatively, is there a similar rank condition for Parrilo's sum-of-squares relaxation?
Relevant answer
Answer
Dear
I don't understand ur question. Please explain it more. I suggest u that u should read my research paper by following me. Because my research work similar with your problem. May be it will helpful to solve ur problem.
  • asked a question related to Convex Optimization
Question
4 answers
I want to know what will be the dual of the problem in attached file. 
Relevant answer
Answer
Check the derivation (Conic one):
There are typos in your formulation.
  • asked a question related to Convex Optimization
Question
7 answers
I am new to this field of optimization, if someone could suggest some materials to study convex optimization, that would be better. I have started reading S Boyd. But the geometric interpretation of all the things like convex function, affine function, cone, convex hull, conic hull etc are not very clear with that.
Relevant answer
Answer
1. Imagine that there is an inland in the sea. You lead a expedition to the island. The only task of the expedition is to put the national flag on the top of the island. If the shape of the island is convex, than the only thing what you should do is to climb higher and higher. When there is no higher point around you, then it is sure that you are on the top of the island. Thus, put the flag there. However, if the shape of the island is not convex, the same method doesn't work necessarily and you must measure the altitudes of the different peaks.
2. The whole thing can be explained even by an "inside" version. Imagine that you must determine the highest point of a cave. If the shape of the cave is convex, then you see the peak point of the cave from any other point of the cave. It is not true from the non-convex case.
****
In both cases we use the "good" property of the objective function. It is not that in any direction the objective function  always either increases or decreases. The key property is that the way of changing the objective function is always decreasing. It means that if you go ahead 1m meter in horizontal direction, then the increase on the first one meter is greater than on the second one meter, etc. This property determines the curving of the island or the cave. However, to get a convex optimization problem, the set of feasible solution, that is the shape of plan of the island/cave must be convex, too.
  • asked a question related to Convex Optimization
Question
3 answers
I have to design a controller using Convex Optimization and Sum of Squares technique, please suggest some papers on this. What are the advantages of using this techniques from the classical controllers or other robust controllers.
Relevant answer
Answer
I guess the Convex Optimization by Stephen Boyd is a good candidate.
here is the link.
  • asked a question related to Convex Optimization
Question
12 answers
does it mean that the two functions are not antagonists? I have solved the optimization problem using the weighted sum method in MATLAB . 
thank you
Relevant answer
Answer
I guess you are minimizing. My impression is, first of all, that you have dominated (inferior) solutions, so it is nor really a PAreto front, but a set of solutions. For a Pareto front, you should delete all dominated solutions.
Second, although the fron looks straight, my guess is that it is not, and, as usual, it is convex. You are just finding the vertices of the feasible region in the objective space.  
If is were straight, there would be no way to find so many solutions...
I hope it helps.
Cheers,
Vlad
  • asked a question related to Convex Optimization
Question
3 answers
Solving large multi-stage stochastic problems may become intractable. There are some applications permit to solve such problems, among which "non-anticaptivity principle", "decomposition techniques", "lagrangian relaxation", and lately "optimal condition decomposition" are the most known ones. In this regard, what are the advantages and disadvantages of the "scenario reduction techniques" attribute to mentioned techniques to reduce computational burden as well as complexity with a good approximation to the original stochastic optimization problem? In advance, I'd appreciate your supportive message.
Relevant answer
Answer
Dear Carvalho, thanks much for the hints...
  • asked a question related to Convex Optimization
Question
3 answers
Is there a definition of convex optimization problem for COMLEX-VALUED MATRIX VARIABLES where objective and constraint functions are real-valued ?
Are KKT conditions true for convex optimization problem with COMPLEX-VALUED MATRIX VARIABLES where objective and constraint functions are real-valued ?
They are often seen in MIMO communication problems.
Thank you very much.
Relevant answer
Answer
Many thanks for your answers.
  • asked a question related to Convex Optimization
Question
1 answer
I have a system of n atoms and I am using a nonlinear conjugate gradient (Polak–Ribière) to minimize its energy, which is the Lennard Jones potential. I use strong Wolfe conditions for the line search. The code works great and fast for small systems (up to 60 atoms). But for larger systems the method gets very slow due to round off errors. The step size found using strong Wolfe conditions is small and so for large systems each atom must move a very short distance (because this small step is multiplied by atoms positions then added to previous position vector for the whole system) and on a computer the system does not feel it. 
  • asked a question related to Convex Optimization
Question
7 answers
I want to maximize a unknown N-dimensional function f(x1,x2,...xN
each xi is independent each other, xi>0 for all xi, and f is smooth. 
My function is not expensive, so I tried f(x1+dx1,x2+dx2,...)  with randomly generated dx1,dx2...
However, I have to calculate f more than millions of times, and sometimes f stuck in local maxima and doesn't give maximum value.
Is there any faster and simple algorithm that can maximize f?
Thank you in advance.
Relevant answer
Answer
There is no fast method guaranteed to find the global maximum of an arbitrary continuous real-valued function on R^n even if all of its derivatives are known.  From a theoretical point of view you can't do significantly better than doing repeated local search from random starting points.  If you can differentiate your function you can do local search faster by line-searching in the gradient direction for each step in the local search.  That is,  let d be the gradient of f at your current point x.  You search for the maximum of  g(r)=f(x + rd)  for r>0, usually by finding zeros of the derivative of g(r).  Use Newton's method if you can get the 2nd derivative; use the secant approximation otherwise.
   However, in practice your function usually has structure, even if you don't grasp it.  When f is  a concave function plus noise, I've used simulated annealing with great success because the randomization smooths out the little bumps and you climb up the big hill.  Often the heuristic ideas from tabu search, particle swarm optimization, etc. of doing more sampling where you got good answers, and also purposely sampling where you have done no sampling, can do better than random starting points.  Also there is the related idea of fixing a subset of the variables if they have shown up repeatedly at different local optima. 
  • asked a question related to Convex Optimization
Question
4 answers
Specifically, estimation of parameters in complex probability distributions
- Variational Inference. 
- MCMC
Relevant answer
Answer
.
asking for convex and non-convex is close to asking for everything !
.
anyway, when it comes to convex optimization and machine learning, i would recommend
Convex Optimization: Algorithms and Complexity
Sébastien Bubeck
.
regarding non-convex optimization, you might start here
.
  • asked a question related to Convex Optimization
Question
5 answers
Hi,
I am trying to use CVX to implementing the antenna selection problem proposed in the following paper:
A. Dua, K. Medepalli e A. Paulraj, "Receive antenna selection in MIMO systems using convex optimization", IEEE Transactions on Wireless Communications, vol. 5, n.o 9, pp. 2353-2357, 2006.
The problem consists of a channel maximization problem, described as follows:
cvx_begin
    variable x(Mr)
    maximize( log_det( eye(Mr) + EbN0*diag(x)*Hfull*Hfull' )/(log(2)) )
    subject to
        0 <= x <= 1;
        trace(diag(x)) == L;
cvx_end
where Mr is the number of antennas in the full set, Hfull is the full channel matrix and L is the number of antennas to be selected.
The problem is that the CVX is returning 'x' with all elements equal to each other.
Does anyone know why this is happening and how I can solve it?
PS: This is a linear relaxation of the optimum problem, and I am supposed to get the L larger entries of 'x' as the selected antennas.
Thank you.
Regards.
Relevant answer
Answer
Your relaxation looks like a conventional MIMO capacity maximization problem. The output of the problem should depend a lot on what value of Hfull that you are using. Have you tried different values of Hfull?
There is generally no guarantee that diag(x) will have only L dominating values in the optimal solution.
  • asked a question related to Convex Optimization
Question
3 answers
hi dear all; please consider a geometric programming problem such as attached example. this is the cost function of a GP problem. can this function be a convex function without any variable change?
Relevant answer
Answer
What do you mean by "without any variable change"? A variable transformation? 
Well, as Richard states - your rather simple function is strictly convex on the positive  axis.
How does this thing relate to geometric programming?
  • asked a question related to Convex Optimization
Question
7 answers
hi dear all; consider 5 complex nonlinear equations with 5 variables. what is the simplest method for solution?
Relevant answer
Hi. The easiest way is to use  fmincon() function in Matlab. Note that, It may not give you the global optima. 
  • asked a question related to Convex Optimization
Question
9 answers
I have used epsilon constraint method as well ass the sum weighted method to find Pareto point for bi-objective model;in this case I have found same results ,same number of Pareto points,
can we claim which our bi-objective model is not non-convex if the results raised from epsilon constraint method and the sum weight method are same?
Relevant answer
Answer
Then, I think that, in general, just looking at the efficient solutions is not sufficient.  Suppose you have just two continuous variables, x and y, and that your feasible region is defined by the following two quadratic constraints:
x^2  + y^2 <= 1, (x+1)^2 + y^2 >= 1.
This feasible region is non-convex, but if your objectives are max x and max y, then the weighted sum and epsilon-constraint methods will give the same efficient frontier, namely, the region:
(x,y): 0 <= x <= 1, 0 <= y <= 1, x^2 + y^2 = 1.
  • asked a question related to Convex Optimization
Question
8 answers
I need to find a entropy minimizing probability distribution under a convex constraint.
Relevant answer
Answer
The answer, in general, is "No". If you are lucky your set is polyhedron, whence it is enough to compute all the extreme points in the set and then compare function values.
  • asked a question related to Convex Optimization
Question
4 answers
The solution of bi-level model is too hard. Generally, duality, decomposition and evolutionary algorithms are used for solving. 
Relevant answer
Answer
it is absolutely possible, you can check this book on your question:
1- Metaheuristics for Bi-level Optimization
2- Sinha, A., Malo, P., & Deb, K. (2017). Evolutionary Bilevel Optimization: An Introduction and Recent Advances. In Recent Advances in Evolutionary Multi-objective Optimization (pp. 71-103). Springer International Publishing.
for GA, these papers have answered your question:
1- Genetic algorithm based approach to bi-level linear programming
2- A hybrid of genetic algorithm and particle swarm optimization for solving bi-level linear programming problem–A case study on supply chain model
3- I also recommend this work:
A parameterised complexity analysis of bi-level optimisation with evolutionary algorithms, MIT, 2016
Corus, D., Lehre, P. K., Neumann, F., & Pourhassan, M. (2016). A parameterised complexity analysis of bi-level optimisation with evolutionary algorithms. Evolutionary computation, 24(1), 183-203.
  • asked a question related to Convex Optimization
Question
3 answers
i have convex problem that start solved for first 50 iteration the after that its go to inaccurate solved .How to resolve the problem?
Relevant answer
Answer
A remark might be added to this: if you have a convex problem, as you state, WHY do you use heuristics?? If you know that the problem is convex, you probably have an expression on the objective function and the constraints (if any), so you should be able to solve it accurately by convex optimization methods, such as by utilising (sub-)gradients.
  • asked a question related to Convex Optimization
Question
5 answers
Is there any condition (similar to K.K.T condition in convex optimization) can be used to get the analytical solution.
I want to get some theorems form the optimization model, and analyze the relationship between the variables and the parameters.
How about the Multi-modules (used to prove the optimality of the model) ?
Relevant answer
Answer
In the case when integer programming is equivalent to linear programming - such as is the case with integer programs that are equivalent to linear network flow problems (like the shortest paths problem) - the KKT conditions are necessary and sufficient. But for integer programs that cannot be translated into a network flow problem or a 2-matroid problem this is not going to work. 
  • asked a question related to Convex Optimization
Question
10 answers
One optimization problem is called system-centric optimization, which is the objective of the authority.
Another set of optimization problems is called user-centric optimizations, which represents the objectives of the users/ the customers.
For the authority, they of course hope to achieve the system-centric goal. So, the actual final result is based on user-centric optimization. However, the users/customers only hope to maximize their own utility which is somehow contradicted with the system-centric optimization.
However, the authority can adjust his one or more parameters in user-centric optimization problems. How to adjust those parameters so the  user-centric optimization's result/solution is aligned with that of the system-centric optimization? Or how to adjust those parameters so that the objective function of the system-centric optimization problem can be minimized (Suppose it's a minimization problem).
Relevant answer
Answer
Your question is slightly unclear. However, in general, if there are more than one agent involved in optimisation problems and where decisions of the agents affects not only their one objective functions, but also the objective function of other agents; game theory is the tool to turn to. Good luck :)
  • asked a question related to Convex Optimization
  • asked a question related to Convex Optimization
Question
4 answers
What is use of xmax in Quantum integrated Particle Swarm Optimization (QPSO)? How can I define xmax in QPSO? How it signifies QPSO?
Relevant answer
Answer
Sorry, I don't get what actually you mean by 'I am not sure about the coding publicly shared by Mr. Zhifei Li under TSP_QPSO title (whether it is the Quantum-PSO or not).' I am not talking about his codes. Am I?
  • asked a question related to Convex Optimization
Question
3 answers
please consider maximum allowable power PT in an OFDM power allocation problem when Pi is the power value of each subcarrie . can we use equality constraint: SUM(Pi)=PT instead of SUM(Pi)<PT inequality?
Relevant answer
Answer
Depends a bit on what is your optimization problem (i.e., what is the cost/utility and what other constraints are there). You haven't shared this one yet, so we can only guess.
Also, you probably meant SUM(Pi)<=PT instead of <PT (the latter being nonconvex).
If your target is monotonic in the variables Pi (which in power allocation problems is typically the case when the target is some sort of rate / mutual information, at least as long as there is no interference) and there are no other constraints influenced by Pi then usually yes. Here is why: if the sum of the powers is less than PT we can increase any of the Pi without violating the constraint. If your target is monotonic in the Pi this will improve (or at least not worsen) the target. Therefore you can expect the inequality to be tight at the optimum.
However, this is not true in general, it depends on your target function!
  • asked a question related to Convex Optimization
Question
5 answers
i was wondering if there is any approach that can reformulate clustering algorithm by standard optimization and maybe relax it into a standard convex optimization with generalized inequality in proper cones like SDP or SOCP.?
Relevant answer
Answer
Although it does not address generalized inequality in proper cones, I believe the deterministic annealing framework proposed by K. Rose has a potential to be adapted to this problem:
Rose, Kenneth. "Deterministic annealing for clustering, compression, classification, regression, and related optimization problems." Proceedings of the IEEE 86.11 (1998): 2210-2239.
  • asked a question related to Convex Optimization
Question
12 answers
I need to theoretically prove the convergence rate of my optimization algorithm. Please suggest some review papers or materials to understand the concept of convergence rate, its different types and process of derivation of convergence rate.
Relevant answer
Answer
For some reason my second answer didn't show up.
I was going to recommend Numerical Optimizarion by Nocedal and Wright http://home.agh.edu.pl/~pba/pdfdoc/Numerical_Optimization.pdf 
You also can check their webpages to find more updated research
  • asked a question related to Convex Optimization
Question
6 answers
I know that there are elegant formulas for the KKT conditions of conic optimization problems with convex objective function for particular cones (such as for example the direct product of Lorentz cones). Are there similar formulas for a general cone?
Relevant answer
Answer
you mean; Nonlinear Optimization by Andrzej P. Ruszczynski, Princeton University Press (1 Jan. 2006
  • asked a question related to Convex Optimization
Question
8 answers
How to find the trade-off point when both objective functions in a multi-objective optimization problem are subjected to maximization, while the obtained Pareto front become concave up, decreasing (like what is common in minimization problems) ?
Relevant answer
Answer
Iman, 
I believe what you are referring to is the Pareto Compromise solution (the trade-off point in the 1st Pareto Frontier). 
If that's the case, please follow the link below to one of the best literature on that topic. 
I used this model for a multi-objective "minimization" problem, yet I believe if you understand the mathematics behind it, you can utilize it for a maximization problem. 
Best of luck, and let us know how it went please. 
  • asked a question related to Convex Optimization
Question
3 answers
Can you provide me with such an example (even in the finite dimensional setting) ?
Relevant answer
Answer
The outer norm of a convex process T is
|T|=sup_{|x|\leq 1} sup_{y\in F(x)} |y|
  • asked a question related to Convex Optimization
Question
5 answers
a/a/a/a/a/a/
Relevant answer
Answer
I really do not understand your problem statement. You first speak about Lagrangian relaxation applied to alinear problem but your problem is an MST and then when asked by Hossein you say that your problem is nonlinear. Can you please make a real statement of your true problem? Otherwise we cannot help you but only confuse you even more.
  • asked a question related to Convex Optimization
Question
1 answer
There are forward-backward, Douglas-Rachford, and primal-dual schemes in the splitting methods family.  But in the literature, it is very rare to see the derivation of these methods.  The authors usually give fancy and sometimes elegant iterative expressions followed by the proof that the proposed method converges.
It seems that the root of splitting methods are from the subgradient of f(x)+g(x) contains zero.  But it also seems that there are many ways after that to get a fixed point.  In addition, there are more general splitting methods that handle the summation of three terms or more.
Could anyone give some suggestions and point out a reference?  Thanks.
Relevant answer
Answer
There are quite many expositions on the subject. Indeed several are quite technical. I quite like the initial sections of the PhD thesis of Jonathan Eckstein (from MIT) - available at 
It provides quite many links between fields, and derives the various classic splitting methods in a natural and not too technical way. While the thesis is from 1989 I think the main issues and building blocks of iterative algorithms are found in there. 
Of course the problem does not have to correspond to the optimality conditions of a convex optimization problem - it is much more general than that, and includes variational inequality problems, saddle-point problems, and more.
  • asked a question related to Convex Optimization
Question
4 answers
Hello;
I have ended up with a quadratic function of X in the form of:
f(X)= XTAX-XTBX+CTX with A and B as positive definite matrices. Note that the second term has a negative sign. Clearly the function is not necessarily convex, but based on some experimentation and prior calculations, it should convex. I wonder to know if there is any analytical or famous experimental method to prove the convexity of this quadratic function.
Thank you
Relevant answer
Answer
i agree with Octav;
f is convex iff A-B is positive semidefinite,
f is strictly convex iff A-B is positive definite,
f is  concave iff A-B is negative semidefinite,
f is strictly concave iff A-B is negative definite.
Greetings, Tadeusz
  • asked a question related to Convex Optimization
Question
23 answers
I have formulated optimization problem for building, where cost concerns with energy consumption and constraints are related to hardware limits and model of building. To solve this formulation, I need to know if problem is convex or non-convex, to select appropriate tool to solve the same. 
Thanks a million in advance. 
Relevant answer
Answer
As pointed out the basic idea is to check if the objective function is convex (for minimization problems) AND the feasible region is convex. There are some ways to check it. Analytically, you can apply the definitions of convex function and set (epigraph can help to establish a relationship between them). You can also try to establish if a complicated function is convex by starting from a simpler convex function and obtain yours by means of operations which preserve convexity. All this could be difficult if you have a very complicated function, though. Depending on the problem, you may also try to solve it with a solver which give you just a local optimum if the problem is non convex (as ipopt, or bonmin if you have integer variables) and then compare the solution with that of a global MINLP solver as Couenne/BARON/SCIP. If the results are very different the problem is probably non convex. It could also be that some solver gives you as output the information about the convexity of the problem, but I am not sure about that.
  • asked a question related to Convex Optimization
Question
1 answer
I am looking for convex optimization solver based on Hadoop MapReduce. I am also interested to convex optimization in a Big Data context.
Best regards,
Sabeur
Relevant answer
Answer
Marc Suchard and colleagues have worked with parallel computation for a number of years.  In the following, they give a characterization of when you can use parallel methods and closely related GPU methods.
Understanding GPU programming for statistical computation: Studies in massively parallel massive mixturesMA Suchard, Q Wang, C Chan, J Frelinger, A Cron, M West
Journal of Computational and Graphical Statistics 19 (2), 419-438
In general you cannot use iterative-fitting-optimization methods because they must make multiple passes over all of the data.  Hadoop Map Reduce are also not good for certain search-retrieval algorithms.
  • asked a question related to Convex Optimization
Question
16 answers
I am working on a part of a paper related to topological properties of boundary points. It is important for me to realize the topological and algebraic behavior the boundary points of the convex sets. I would be grateful if someone could help me around this issues by giving some ideas or references related to it.
The general question provided in following;
Let B be a closed set in n−dimensional Euclidean space. What other properties B should have in order to be guaranteed that there exist the closed convex set A such that ∂A=B. How about infinite dimensional spaces?
Relevant answer
Answer
A discussion on this question is included at the page 285 of the recent book by Valeriu Soltan, Lectures on Convex Sets, World Scientific 2015.
  • asked a question related to Convex Optimization
Question
14 answers
The problem is: min f(x),
s.t: Ax>=b,
where f is convex.
Relevant answer
Answer
CPLEX, Gurobi, XPRESS and Mosek all have strong convex QP solvers that are well suited to solving your model to optimality, either with the barrier method as Adam mentioned or the primal or dual simplex methods.   If you have access to one of those solvers, give them a try.
Also, you indicated that you just wanted a feasible solution.   In that sense, in the unlikely event that if you found the QP was hard to solve to optimality for the aforementioned solvers, you could try some alternate objective function that is easier to handle.   Possible alternate objectives include:
-  The zero objective.   Of course, this pays no attention to your objective function, but it might jump start the subsequent quadratic optimization.
-  Instead of minimizing the 2 norm of x, minimize the one norm, i.e. the sum of the absolute values of each element of x.   This can be modelled as a linear program; you just replace each x element that appears in an absolute value with the difference of two nonnegative variables.   As long as you solve the LP to an optimal basic solution, one of each pair of nonnegative variables must be 0, and you can express the absolute values as the sum of the two nonnegative variables in the pair.
- Use a piecewise linear approximation to your QP objective.   Since it is convex, this will result in an LP as well.
Using alternate objectives probably is unnecessary for a continuous QP, but it can be very helpful when solving a convex (or nonconvex) QP with integers restrictions on the variables.
  • asked a question related to Convex Optimization
Question
1 answer
In practice, how does one normally detect if the problem is in-feasible when optimizing using ADMM methods ?
Relevant answer
Answer
In " Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers ", it is suggested to put a threshold on the primal and dual residuals. I hope this answers your first question. 
  • asked a question related to Convex Optimization
Question
11 answers
I need to comment whether my optimization function is convex or non-convex. My optimization function is in the form of (y-y_cap)^2. y is know. y_cap comes out of a MATLAB pfile. So, y_cap can be considered output of a black-box model. It is also not possible to plot y-y_cap as it is a 22*8 matrix
Relevant answer
Answer
Please think carefully, when you pose a question. Your question is extremely unclear, and this  despite the fact, that mathematics enables you to pose questions clearly!
  • asked a question related to Convex Optimization
Question
8 answers
In convex optimization, what is the definition of the facet of a convex hull, and how to prove a constraint is facet-defining? Can anyone recommend some related references?
Relevant answer
Answer
An answer to this comes in chapter 9 of Wolsey's book on Integer Programming around page 142. The full reference is: Wolsey, L. A. (1998). Integer programming (Vol. 42). New York: Wiley.
  • asked a question related to Convex Optimization
Question
14 answers
For a given pdf function f(x) defined in [0,1], which is monotonically increasing and convex, I need to find a probability generation (polynomial) function Q(x) of a maximum degree D, where its derivative is always larger than or equal to the pdf function in [0,1-e], for every e>0, Q'(x)>f(x). It can be solved by numerical optimization algorithms, but I am looking for some theorems to prove the existence of such a PGF for every e.
Please find here more details on the question:
Relevant answer
Answer
1. The derivative of a polynomial of degree D is a polynomial of degree D-1, so you are looking for a polynomial (with bounded degree by a prespecified bound) that dominates your pdf. 
2. Dominating f on [0,1-e] for every e>0 is the same as dominating f on [0,1). 
3. As your f is convex, what is wrong with using the cord (a linear function = polynomial of degree (at most) 1) stretching from the point (0, f(0)) to (1, f(1)) on the graph of f ? 
  • asked a question related to Convex Optimization
Question
7 answers
I want to use the SOSTOOLS toolbox, which works in Matlab and can be combined with the following SDP solvers: CSDP, SDPNAL, SDPASeDuMi, and SDPT3. I'm not sure about which of these solvers I should choose. Any advice would be welcome.
Relevant answer
Answer
SDPT3 is more numerically stable compared to SeDuMi.
  • asked a question related to Convex Optimization
Question
4 answers
I would really appreciate if anyone can give me a formal citation for this. I know this is a standard problem for which I also have a proof. However this is requried as a part of a more complicated problem and it would be nicer for the writing if I can just simply give a citation.
Relevant answer
Answer
For the same question  see 
and discussion there. It turns out that it is an interesting question  which may related to the  max -min theory.
Now we will outline a proof.
Perhaps, we can suppose  that  X   and Y are metric spaces (in particular   subset of real line R)  and that f(x,y)    is real-valued function   in x ,y.   For a simple model we can   take X=R  and Y=[a,b] closed interval  in R.
Set  I(x) = inf_{y\in Y} f(x,y).   Fix a point  x_0 in    X .
Let us prove that   I(x)    is continuous at x_0.
 Consider a sequence x_n   in X   such that   x_n    converges to  x_0.
Let us prove that    I(x_n) converges to I(x_0).
Since Y is compact   there is a  sequence o  (y_n)    and  y_0 in Y   such that   I(x_n) = f(x_n,y_n)  and   a_0:= I(x_0= f(x_0,y_0).   Since Y is compact there is a subsequence of y_n   which converges to y^0  in Y.  By abusing notation we can suppose that  (x_n,y_n) converges to  (x_0, y^0).  By continuity of f in x ,y    we conclude that  f(x_n,y_n)  converges to   f(x_0, y^0):=b_0.  By definition of I(x_n),        f(x_n,y_n) <= f(x_n,y_0)  and therefore     since   (x_n,y_n)   converges to (x_0, y^0)   and    (x_n,y_0)    converges to (x_0, y_0), by continyity of f,    we find
b_0=f(x_0,y^0)<= a_0.  On the other hand,  by definition   a_0<= b_0. Hence a_0=b_0.
This shows that the set of  cluster points   (or accumulation points) of sequence
I(x_n)  is a  single point a_0  and therfore   I(x_n)  converges to  a_0=I(x_0).
Therefore  I  is continuous in x_0 .
If necessary, we can give more details.
You can also see Mathematical Analyis I/II. by Vladimir A. Zorich.
''An entire generation of mathematicians has grown up during the time - tween the appearance of the first edition of this textbook and the publication of the fourth edition, a translation of which is before you. The book is familiar to many people, who either attended the lectures on which it is based or studied out of it, and who now teach others in universities all over the world. I am glad that it has become accessible to English-speaking readers. This textbook consists of two parts. It is aimed primarily at university students and teachers specializing in mathematics and natural sciences, and at all those who wish to see both the rigorous mathematical theory and examples of its effective use in the solution of real problems of natural science. The textbook exposes classical analysis as it is today, as an integral part of Mathematics in its interrelations with other modern mathematical courses such as algebra, differential geometry, differential equations, complex and functional analysis''.
  • asked a question related to Convex Optimization
Question
1 answer
There are efficient methods in DC programming. On the other hand every C1 function with a Lipschitz derivative on a compact convex set is the difference of a convex function and a convex quadratic function, e.g., J. Glob. Opt. 46(2010)155-161. This means that one can optimize C1 functions by DC methods. Has anybody tried this approach?
Relevant answer
Answer
The question can be made clearer: Namely, every C1 function with a Lipschitz derivative on a compact convex set can be made (strictly) convex by adding to it a simple quadratic function. Illustration: sin x + x^2 is strictly convex. This yields a possibly non-intuitive result: Every such C1 is the difference of two convex functions. For such differences there exist efficient methods in DC programming (D = difference, C = convex functions). The question is whether anybody has tried this approach to optimize a C1 function. (The problem of characterizing DC functions, i.e., differences of convex functions seems to be a non-trivial problem but one does not have to go there.)
  • asked a question related to Convex Optimization
Question
10 answers
As a part of my work I need to obtain all the extreme optimal solutions of a given LP. So far I came up with the solution where I find the value of the LP, fix it as a constraint and find all the feasible extreme solutions of the resulting LP, which leads to finding all the vertices of the polyhedron formed by the constraints of this LP. Can somebody please direct my to some work where there is a worst case complexity analysis of finding all vertices of such polyhedron?
Relevant answer
Answer
The optimal face of a polyhedron is a polyhedron itself. As a consequence, its number of vertices may grow exponentially with the input size. If the optimal face is, e.g., an hypercube with dimension d, you can descibe it by 2d inequalities in d variables, but its vertex set ha cardinality 2^d.  The classic book by Katta Murty on Linear programming is a source of insights on these kind of issues.
  • asked a question related to Convex Optimization
Question
42 answers
ABSOLUTE VALUE OPERATOR LINEARIZATION
I have a nonlinear term in the objective function of my optimization problem as an Absolute Value function like |x-a|.
As far as I know, an Absolute Value operator makes the optimization problems nonlinear (i.e. NLP). How can I make it linear (LP or MILP)?
Max  f(x)=g(x) + b*|x-a|
   s.t.   some linear constraints
Regards,
Morteza Shabanzadeh
Relevant answer
Answer
max f(x)=g(x)+b*p+b*q
s.t
other constraints
x-a+p-q=0;
p,q>=0;
  • asked a question related to Convex Optimization
Question
16 answers
Binary Variable * Real Variable = ?
1) lead to an equivalent 'Nonlinear' variable (and thus => MINLP),
2) lead to an equivalent 'Integer' variable, 'Discrete' I mean (and thus => MILP).
Which one is correct and why?
What is your idea to deal with this problem by adding a constraint and make the resultant problem MILP (if it is not MILP).
Regards,
Morteza Shabanzadeh
Relevant answer
Answer
Your product just tries to express that the continuous variable should be zero if the binary variable is zero.
In the Framework of a MILP you could as well avoid the product by generalized upper bounds on the continuous variable - if it has a natural upper bound -- AND if the variable does not occur elsewhere without that product.
In that case just add the two inequalities
binary * lower_bound <= continuous <= binary * upper_bound
to the problem and your variable is forced to zero if binary is zero.
  • asked a question related to Convex Optimization
Question
8 answers
Since the strict complementarity condition (between the Lagrange multipliers and the inequality constraints) is not guaranteed for the optimal solution of a quadratic programming problem, I wonder if how I can have the closest solution to the optimal one which holds the strict complementarity condition.
Relevant answer
Answer
As you can see from your own example, there exists no "nearest" or "closest" solution. That would be equivalent to asking what the projection is on an open set - there is an optimal least value (0) but no solution, since however close you produce a point I can find one even closer by taking the mean value of the "open" boundary and the point you give me. 
It's a problem without a solution. IF I have understood you right. :-) Otherwise, I'm wrong! :-) 
  • asked a question related to Convex Optimization
Question
1 answer
I have an optimisation problem with a two-stage cost function, consisting of two parts. For a parameter value less than a threshold (which is an optimal value by itself), only the first part appears in the optimal value function. For the parameter value more than this threshold, on the other hand, both parts appear in the optimal value function.
I tried to under-estimate the problem by under-estimating all non-convex terms in the cost function and constraints; however, the optimal value function of the resulting under-estimated problem is not convex because of the two-stage operation.
I am wondering if there is any way to under-estimate such a problem considering the multi-stage operation in order to have a convex optimal value function of the resulting under-estimated problem.
Relevant answer
Answer
Thanks Isaac. The problem is that the 'threshold parameter' is not known beforehand and, indeed, must be found by solving the problem.
Therefore, if I use two binary variables, as you suggested, I think I still need to under-estimate the similar problem again in order to solve the problem for the case that e.g. both binary variables are '1'. 
More comments would be appreciated.
  • asked a question related to Convex Optimization
Question
5 answers
How do you minimize the following non linear optimization problem:
find x* that minimizes the H-infinity-norm: || G(s, x) ||oo
Subject to : xmin < x < xmax
Where:
s=j*w and w = wmin to wmax { for example: w = logspace (-3,+3,50)}
G: denotes the transfer matrix that given (for example as):
G(s,x)=[(s^x1)+3/s -9+x2*s+x3; 1+x1^x2 s/x3]
x=[x1;x2;x3] denotes the design parameter vector to be determined by an optimization method
Relevant answer
Answer
Dear Hadi. First, i you thank's for your response.
I'm previously use the PSO algorithm but for the real and simple objective functions. Now, our problem is that to evaluate the H-infinity norm of the transfer matrix G(s,x).
I can make the objective function in Matlab as follow:
Function  norm_G=myfunction(x)
w=logspace(wmin,wmax,N)
  for m=1:length(w)
         j=sqrt(-1);s=j*w(m);
         G=[(s^x(1))+3/s -9+x(2)*s+x(3); 1+x(1)^x(2) s/x(3)];
Now how to evaluate the H-infinity norm of G
end
  • asked a question related to Convex Optimization
Question
10 answers
in convex optimization who can explain what wi will say by this :
minimize 1/2 xTAx+cTx
subject to Dx=b
0<=x<=d
?
Relevant answer
Answer
The feasible set {x : Dx=b and 0<=x<=d} is a polytope (here obtained by intersecting a finite number of hyperplanes and an n-dimensional box). So, you are trying to find the minimal value (and a point at which this minimum is attained) of a quadratic function q(x) := 1/2x^TAx + c^Tx, on a polytope.
Hope this helps :)
  • asked a question related to Convex Optimization
Question
8 answers
Software to linear programming as cplex, gurobi, etc but with the method of aggregation of columns integrated(Dantzing wolfe)
Is possible save memory in big problems with this method?
Relevant answer
Answer
thanks, I not know this software
  • asked a question related to Convex Optimization
Question
5 answers
Thanks in advance.
Relevant answer
Answer
This is a good question with more than one answer.
A good place to start answering this question is
S. Wasowicz, Support-type propeties of generalized convex functions, J. Math. Anal. Appl. 365, 2010, 415-427:
A direct  generalization of higher-order convexity appears to be convexity with respect to Chebyshev systems (p. 1).   Cheyshev systems are introduced in Section 2.2, starting on page 3.   Generalization of convexity is considered in Section 2.3, starting on page 4.
Another good place to consider is
M. Bessenyei, Hermit-Hadamard-type inequalities for generalized convex functions, Ph.D. thesis, University of Debrecen, Hungary, 2004:
This thesis is especially good inasmuch as it considers generalized 2-convexity in Chapter 2, starting on page 25.   It is also considers generalized convexity induced by Chebyshev systems (see Chapter  3, starting on page 41).
Generalized ocnvexity cones are introduced in
Z. Ziegler, Generalized convexity cones, Pacific J of Math. 17, 1966, no. 3, 561-580:
Although this is an early paper, it is still excellent and well-worth considering.
  • asked a question related to Convex Optimization
Question
2 answers
Hello,
For those of you who are familiar with Pincus Theorem which is based on finding a global maximum of ANY objective function. 
Can anyone, kindly, provide me with another similar theorem which finds the global maximum of any objective function (No constraints here).
Relevant answer
Answer
I hope this can be helpful: this paper deals with a similar problem, and indeed, cites Pincus's Theorem and compares its main results with it:
Conditions for Global Optimality in Nonlinear Programming
James E. Falk
Operations Research
Vol. 21, No. 1, Mathematical Programming and Its Applications (Jan. - Feb., 1973), pp. 337-340
  • asked a question related to Convex Optimization
Question
8 answers
Consider the following setting: g(y,V) = minw\in C(y-w)TV(y-w), C is a closed convex cone, V is positive definite and y is in Rn.
Assume an appropriate norm for (y,V), would g(y,V) be continuous in (y,V) ?
I know if C is compact then this is easy to prove. But a cone cannot be compact.
At first glance I thought this should be a trivial problem but whatever strategy I tried I got stuck in the compactness of C. So I turn to the literature and it turns out so far that continuity of such a function like g(y,V) seems to be an active area of research ( I do hope I am wrong!)
I am very naiive to convex optimization in general, particularly conic projection. So I will appreciate any hints or comments on how to tackle this problem as well as solid disproof!
Many thanks in advance
Relevant answer
Answer
Well, let's look at it this way. If V is a real-valued, symmetric and positive definite matrix, then your objective function is strongly convex on R^n. Such functions have a unique minimum over any non-empty, closed, and convex set in R^n. It appears that your problem fits within this framework. Danskin's Theorem of course talks about a strictly convex objective function over a closed, convex and COMPACT set. So here, you want to replace the compactness with something else, and the uniform strong convexity of your objective function ensures just that. 
By the way, here is a link to a revealing note on the issue:
  • asked a question related to Convex Optimization
Question
6 answers
For g(y,V,w)=(y−w)TV(y−w), for y in Rn, V is positive definite (e.g. variance matrix) and w is in a closed convex cone C in Rn.  
Obviously for every fixed y and V minw\in C g(y,V,w) exists
But my question is for every fixed y,V and y0, V0 does
minw\in C [ g(y,V,w) - g(y0,V0,w) ] exist?
Relevant answer
Answer
Did you really mean to have a w_0? If so I'm not sure how it is to be regarded.
Assuming not, suppose you work in one dimensional space. Take V and V_0 to both be the matrix ((1)). Take y=0 and y_0 = 1. So the functions are x^2 and (x-1)^2. Their difference can be arbitrarily large or small depending on the convex cone region specified (a ray, in this case).
  • asked a question related to Convex Optimization
Question
6 answers
Hello,
I would like to know that how I can find the number of variables (especially the integer ones) in GAMS (General Algebraic Modeling System) codes.
Does GAMS platform have any options to show the number of variables?
Any help would be appreciated.
Regards,
Morteza
Relevant answer
Answer
Hi,
If you are working with GAMS IDE (the Integrated development Environment of GAMS that runs in Windows), then you can find the number of variables (and their type) in the log file. The log file is the window that pops up when you run your code. You can look for the number of columns (i.e. the variables) and the number of integer-columns (i.e. the number of integer variables)
That is the easiest way for me to identify the number of variables, but maybe someone else has a better way to do it.
Regards,
Laura
  • asked a question related to Convex Optimization
Question
6 answers
For example, if I have in matlab matrix of form [A*B;A*C;A*D] how I can check if it represents the result of Kronecker product?
Relevant answer
Answer
More naively: if F = kron[F1,F2], where F is m by n and F2 is d by e, then F is partitioned into blocks, each of which is a constant multiple of F2. (The constants are the entries in F1.) This means d divides m and e divides n. So for each such pair (d, e), excluding the trivial pair (1,1), 
  • Partition your matrix into d by e blocks, and choose and fix one such block whose entries are not all zero (to whatever precision you use).
  • For each of the other blocks, compare each entry with the corresponding entry in the original block. If the original entry was zero, then the new entry must be zero. If not, then  divide the new entry by the original one: the ratio must be the same throughout the new block.
It should be easy to program and quick to run, because failure with a single entry will exclude that pair (d,e), and the number of divisors grows slowly (and also oscillates wildly).
  • asked a question related to Convex Optimization
Question
201 answers
In a Euclidean space, an object S is convex, provided the line segment connecting each pair of points in S is also within S. Examples of convex objects in the attached image include convex polyhedra and tilings containing convex polygons.  Can other tilings containing convex shapes be found?
Solid cubes (not hollow cubes or cubes with dents in them) are also examples of convex objects.   However, crescent shapes (a partial point-filled circular disk) are non-convex .   To test the non-convexity of a crescent, select a pair of points along the inner edge of a crescent and draw a line segment between the selected points.   Except for the end points, the remaining points in the line segment will not be within the crescent.  Except for the 3rd and 5th cubes, the cubes in the attached images are convex objects (all points bounded by walls of each cube are contained in the cube).
From left-to-right, the cresent shapes are shown in the attached image are non-convex: Nakhchivan, Azerbaijan dome, Taj Mahal, flags of Algeria, Tunisia, Turkey and Turkmenistan. For more examples of crescent objects, see
Can you identify other crescent shapes in art or in architecture that are non-convex?  Going further, can you identify other non-convex objects in art or in architecture?
The notion of convexity leads to many practical applications such as optimization
image processing
and antismatroids, useful in discrete event simulation, AI planning, and feasible states of learners:
In science, convex sets provide a basis solving optimization and duality problems, e.g.,
Convex sets also appear in solving force closure in robotic grasping, e.g.,
Recent work has been done on decomposing 2D and 3D models into their approximate convex components. See, for example, the attached decompositions from page 6 in
J.-M. Lien, Approximate convex decomposition and its applications, Ph.D. thesis, Texas A&M University, 2006:
There are many other applications of the notion of convexity in Science. Can you suggest any?
Relevant answer
Answer
It is always lovely when you can form a convex polyhedron with a discrete optimization problem.  It gives you properties you can exploit, some that you have pointed out.  The application I'd like to point out is Scheduling.  Scheduling is pretty notorious for exploiting properties of convex polyhedra if the problem as an optimization problem can be formulated as a relaxed LP (of its IP counterpart).  Of course this usually means solving a linear program then rounding can be used (or in some realms is avoided, but instead primal-dual approaches are found at times).  
Here is an example of how (the rounding theorem in this paper is the crux of how):  https://www.researchgate.net/publication/4355176_Approximation_algorithms_for_scheduling_unrelated_parallel_machines
The above link is to a very important 2-approximation algorithm that showed how you can tackle heterogeneous computer scheduling (unrelated parallel machines) with linear programming, and still matches the best approximation factor to this date (some others have been proposed with the same approximation factor).  New results still draw from this paper to this date (see the cited in, there are a lot).
The notion of convexity can be applied to optimization problems and give us a better understanding of how well or just how we can approximate intractable problems.
Hope this helps :)!