Science topic
Convex Optimization - Science topic
Explore the latest questions and answers in Convex Optimization, and find Convex Optimization experts.
Questions related to Convex Optimization
Schauder Fixed Point conjecture deals with the existence of fixed points for certain types of operators on Banach spaces. It suggests that every non-expansive mapping of a non-empty convex, weakly compact subset of a Banach space into itself has a fixed point. The status of this conjecture may depend on the specific assumptions and settings.
The steepest descent method proposed by Fliege et al. motivates the research on descent methods for multiobjective optimization, which has received increasing attention in recent years. In the context of convex cases, all the Pareto critical point can be obtained by weighted scalarization method with cheap iterates. Therefore, is it necessary to develop descent methods for convex multiobjective optimization problems?
linear least squares problem::
fit f(t) = sin(t) on the interval [0, π] with polynomials of degree n with n = 1, 2, 3, · · · , 10. Here, we use equally spaced nodes.
Solve the normal equation with a backslash in Matlab. Save the norm of the residual and the condition numbers of AT A??
Could anybody please tell me how can find x, y, and A in that case ??
Dear Scholars,
I would like to solve a Fluid Mechanic optimization problem that requires the implementation of an optimization algorithm together with Artificial Neural network. I had some questions about Convex optimization algorithm and I would appreciate it if you could give some advice to me.
My question is about possibility of implementing of the Convex optimization together with Artificial Neural Network to find a unique solution for a multi-objective optimization problem. The optimization problem that I am trying to code is explained as the following equations. The objective function utilized in the optimization problem is defined in the following equation:
📷
Where OF is the objective function, wi are the weights assigned to each cost function, Oi is the ith cost function defined as the relative difference between the experimental and the predicted evaporation metrics for fuel droplet (denoted by superscript of exp and mdl, respectively), k is the number of cost functions (k = 4, equal to the number of evaporation metrics), c [c1, c2, and c3] is the mass fraction vector defining the blend of three components, subjected to the following constrains:
📷
Due to high CPU time required for modeling and calculating the objective functions (OF), an ANN was trained based on some tabulated data from modeling of fuel droplet evaporation and used for calculating the OF through optimization iteration.
In the same manner, the wi values are subjected to optimization during the minimization of OF, with the following constraints:
📷
It worth mentioning that, I have solved this problem by employing Genetic Algorithm together with ANN. Although, the iterative process for solving the problem converged to acceptable solutions. But, the algorithm did not return a unique solution.
Regarding that, I would ask you about possibility of using Convex optimization algorithm together with ANN for solving the aforementioned problem to achieve a unique solution. In case of such feasibility, I would appreciate it if you could mention some relevant publications.

For example :
structures & aerodynamics disciplines.Objective functions: weight, wing displacement and aerodynamic coefficient.
The question is about finding convexity of any optimisation problem involving equality constraints. I will start with my belief first so if there are already some issues there, they can be addressed right away.
Belief: A problem is convex iff
1. Functions defining inequality constraints i.e. g_i(x) <=0 are all convex. The easiest way to check this is by finding hessian of all g_i(x) and it should be positive semi-definite (all eigen values must be non-negative).
2. Functions defining equality constraints i.e. h_i(x) = 0 must be all linear.
3. The objective function must be convex "over the feasible set".
Now the issue I am facing is with the 3rd one. I believe that the function does not necessarily need to be convex over R but only over the feasible set defined by g_i(x)<=0 and h_i(x)=0. Now, if we would just have inequality constraints, it would be easy to determine whether objective function is convex or not. We can just find it's hessian and see whether the hessian stays positive semi-definite for all values within the feasible set (If it would be a quadratic function, the hessian would be constant and the values of X will not matter anyway). BUT it is a bit difficult to do this exercise with the equality constraints like how do we test the hessian for values of X which satisfy these equalities.
I will highlight this issue with a simple example:
Let's say we have a problem with objective function
f(X) = -x*y with X = [x;y]
and equality constraint
h_(X)=x+y =0
Now the equality constraint is linear so the feasible set is convex. f(X) however seems to be non-convex at first glance as if we find hessian(f,X), it comes out to be [0,-1;-1,0] which has eigen values as -1 and 1. Thus, the hessian is not positive semi-definite and f(X) is not convex (at least not over R).
On the other hand, if we would choose to eliminate y by using h_(X) i.e. y = -x, we get an f(x) = x^2 which is totally convex over R1. Does this mean that f(X) is convex over the feasible set defined by x+y=0??
If yes, then how should we check the convexity of the objective function when equality constraints are involved? We clearly see that just checking the hessian of f(X) may not be enough!

I am interested in global convergence and application of that algorithm to different area.
Any help is highly appreciated
Hello, I have a basic question,
I know that the convex optimization problem is the one in which its objective and inequality functions are convex and its equality function is affine. My question is which set the objective and inequality functions have to be convex on? Is it their entire domain? or just the feasible set of the problem?
For example: let the problem be minimizing f0(x),
s.t.:
fi(x)<=0. for i = 1, 2, ..., m.
and f1(x), for example, is not a convex function on its entire domain, but it is convex on the feasible set of the problem. Is this problem is meant to be convex?
I am trying to minimize the following objective function:
(A-f(B)).^2 + d*TV(B)
A: known 2D image
B: unknown 2D image
f: a non-injective function (e.g. sine function)
d: constant
TV: total variation: sum of the image gradient magnitudes
I am not an expert on these sorts of problems and looking for some hints to start from. Thank you.
Hi,
I have a little knowledge about optimization and I have a simple question. If my convex optimization problem has an inequality constraint that ensures that the optimization variable can't be less than zero and greater than a maximum value and I have an iterative algorithm to solve this problem. My question is, can I handle this constraint with programming by forcing the value of the variable to not be less than zero and larger than the maximum value in each iteration (with the min() and max() functions for example)? or I must handle this constraint by two Lagrangian multipliers?
I need to implement a load scheduling algorithm that involves solving an online optimisation problem from a research paper for my Real time systems course.
This convex optimisation problem is setup through Model predictive control or receding horizon control. This problem involves under 100 decision variables. I am just googling about solvers and it would be great if you can give me more accurate suggestions about specific numerical solvers available in python for solving these kinds of cost functions (quadratic + one norm + infinity norm) and constraints (both equality and inequality). I think its convex (single minimum) because the author says so.
*If you are an expert you dont have to read the contents below just click and look at the first attachment and suggest to me a solver or links to a solver in python, Thanks :)*
Full paper at this link: https://github.com/JosePeeterson/pdf_hosting/blob/master/paper_2_RTS.pdf
Please see the cost function and constraints in the attachment.
**Brief description of the problem:**
We want to schedule tasks (i.e. power allocation) that maximises power allocation from renewable sources, W to Electric vehicle (EV) loads, E that have some timing deadlines, di (due to running out of charge). We should minimize the use of grid energy sources, G and only use it in the worst case when EV load demand cannot be satisfied using renewable sources.
In the objective function the matrix W represents renewable energy/power allocation for task i at time step k. we look into N time steps into the future from the current time step, t.
Say we have a renewable energy forecast chart that predicts the maximum output power at future times like in renewable energy generation forecast graph attached. Matrix G is grid generated power/energy also indexed similarly to W and is available at all times. Total power/energy supplied to load E is W + G. phi (flexibility factor) is the term that implicitly brings W into the cost fucntion. we try to keep phi large so that we can always defer load scheduling to future times if renewable power/energy is not sufficiently available for the time being and maybe available in the future. Note: the size/dimension (MXN) of matrix W and G is decreasing as we approach end of active tasks list in the time horizon N. As tasks in active tasks get completed and as we approach the last deadline in active task list M and N are decreasing.
In constraint 7, we put a cap on the maximum renewable energy available for all the M tasks up to N time steps using the renewable energy generation forcast graph attached.
constraint 8 is a conservation equation W and G power/energy supplied equals load energy required, E for the M tasks.
In constriant 9 mi is maximum power that can be supplied to the load due to loads physical limitation.
In contraint 10 and 11, di is the deadline of task i, delta t is time step for within which the optimisation must be solved. ei(k) is the remanining energy required for the task to complete.
From this description please suggest solvers or links to solvers that can be used in python.
Thank you!!



Mathematical programming is the best optimization tool with many years of strong theoretical background. Also, it is demonstrated that it can solve complex optimization problems on the scale of one million design variables, efficiently. Also, the methods are so reliable! Besides, there is mathematical proof for the existence of the solution and the globality of the optimum.
However, in some cases in which there are discontinuities in the objective function, there would be some problems due to non-differentiable problem. Some methods such as sub-gradients are proposed to solve such problems. However, I cannot find many papers in the state-of-the-art of engineering optimization of discontinuous optimization using mathematical programming. Engineers mostly use metaheuristics for such cases.
Can all problems with discontinuities be solved with mathematical programming? Is it easy to implement sub-gradients for large scale industrial problems? Do they work in non-convex problems?
A simple simple example of such a function is attached here.
What are some of the well-written good references that discusses why finding the penalty parameter when solving a nonlinear constrained optimization problem is hard to find from the computational perspective. What are some of the computational methods done to find the parameter as I understand finding such a parameter is problem-dependant.
Any insights also would be very helpful.
Thanks
Is there such a thing as convex optimization for large deep learning networks? Is there a way to guarantee convergence? Has anyone looked into this?
in the method (HPM) I do not know how to construct the homotopy convex in many papers i find different of type homotopy (is it modification of these method or what ?)
and what is the conditions for construct the correct homotopyconvex
#applied mathematics #Neumerical analysis #Homotopy perturbation method
Considering the powered descent guidance problem for a spacecraft or launch vehicle - Early publications by Acikmese, et.al. used lossless convexification to solve the landing guidance problem using point mass dynamics.
Literature that dealt with the powered descent guidance problem using 6 DoF dynamics, also by Acikmese, et.al. proposed the use of successive convexification, with the reasons being that lossless convexification could not handle non-convex state and some classes of non-convex control constraints.
I have the following questions with respect to the above
1) Why is it that Lossless convexification cannot handle some classes of non-convex control constraints ? and what are those classes ? (e.g. the glideslope constraint)
2) What is it about lossless convexification that makes it unsuitable to operate on non-convex state constraints ,as opposed to successive convexification ?
2) Are there specific rotational formalisms ,e.g. quaternions, DCMs, Dual quaternions, MRPs, etc. that are inherently convex ? if so, how does one go about showing that they are convex ?
I would be much obliged if someone could answer these questions or atleast provide references that would aid in my understanding of the topic of convexification.
Thank you.
I would be grateful if anyone could tell me how the McCormick error can be reduced systematically. In fact, I would like to know how we can efficiently recognize and obtain a tighter relaxation for bi-linear terms when we use McCormick envelopes.
For instance, consider the simple optimization problem below. The results show a big McCormick error! Its MATLAB code is attached.
Min Z = x^2 - x
s.t.
-1 <= x <= 3
(optimal: x* = 0.5 , Z* = -0.25 ; McCormick: x*=2.6!)
I have rewritten an MPC problem to a QP formulation.
The QP solvers quadprog and Gurobi, which uses the interior-point algorithm, give me the same objective function value and optimized values x, but GPAD, a first-order solver, gives me the same optimized values x, but an objective function value, which is a factor 10 bigger compared to quadprog and Gurobi. Do you anyone know a possible explanation?
It seems that the quadprog function of MATLAB, the (conventional) interior-point algorithm, is not fully exploiting the sparsity and structure of the sparse QP formulation based on my results.
In Model Predictive Control, the computational complexity should scale linearly with the prediction horizon N. However, results show that the complexity scales quadratically with the prediction horizon N.
What can be possible explanations?
I am trying to solve a QP problem.
Does anybody know the differences between the interior-point-convex algorithm of quadprog and the barrier method of Gurobi in terms which kind of matrices can the solvers handle the best? Sparse matrices or dense matrices? And what kind of problems: large problems or small problems? Furthermore, which solver needs more iterations (with less expensive cost per iteration) and which solver needs fewer iterations, but expensive cost/time per iteration.
From results, I think Gurobi gives more iterations and quadprog less, but I do not know why?
And what are the differences with GPAD by means of what is described above?
According to first order Taylor series approximation, |xHw|2 can be approximated as:
|xHw|2 >= 2*real{(w0)HxxHw}-|xHw0|2
where x and w is colum vector and w is variable vector which is unknown.
How to find a lower approximate f(w) of |xHw|2*|yHw|2- |zHw|2*|bHw|2 satisfies the follwoing condition:
|xHw|2*|yHw|2- |zHw|2*|bHw|2>= f(w)
where w is variable vector which is unknown, vectors x,y,z,b are known.
Hello everyone,
I am a PhD student with a deep learning background. I am trying to find more robust deep learning models and I just came through defining the problem as a minimax problem as the attacker tries to maximize the loss function and the defender tries to minimize the risk of he attacker. I have just learned that this problem can be solved in three ways : 1- Lower bounds 2- Exact solutions 3- Upper bounds.
The problem is that I do not know how to get deeper in optimization in order to get a unique solution to the problem.
Can you recommend some resources or road-map for optimization, convex optimization, non-convex optimization, and robust optimization?
Hi all,
Is ADMM (Alternating Direction Method of Multipliers) numerically stable? That is, if there is a small error in a problem parameter, would the error in the obtained solution be small as well?
If it is stable, can you showed me a paper that proves the stability? If it is not, do you know any extension to the algorithm that makes it stable?
Thanks,
Alex
I have found many related optimization references, but there were no method or enlightments for my specific optimization problems.
known: D, W, P_max, h_j, T_k, \sigma_u.
Unknown: p_k. k=1,2,3,...K. (for example, K=3 users)
Here, I intend to find the optimization variables of $\mathbf p = [p_1, p_k, ..., p_K]^{T}$。
Hope for some suggestions or enlightments from convex optimization researchers.
Thanks a lot!


I have a relaxed convex problem (Conic) that can be expressed using two formulations. When solved using CPLEX, I obtained the same global solution. However, one formulation yielded the global solution in much faster time than the other. My question is how to explain these finding scientifically? My guess is that the faster algorithm has a smaller feasible space which allow the solver to find the solution in a faster way but I can't prove this or even know if this is really the case. Any suggestion that can help in comparing the performance?
Is linear programming one part of convex optimization?
Deterministic Global optimization relies on convex relaxation of the non-convex problems. Certain nonlinearities are duly converted into linear forms underestimators to be solved by efficient MILP solvers (e.g. signomial functions/ bilinear terms).
Most nonlinearityies are approximated to linear functions by piece-wise linearizations. However, I am wondering if this linearizations guarantees that the approximations are understimators of the original nonconvex problem (i.e. for all x in Domf, f(x) >= u(x) where u is the understimator)
because otherwise the understimator may miss the global optimum during the branch and bound process.
Can the solver still converge even if the relaxation is not an understimator?
min - x - y
s.t. xy <= 4
0<= x <= 4
0 <= y <= 8
Please give their appropriate cases.
The constraints include both linear constraints and nonlinear constraints. The essential issue lies in to how to deal with the nonlinear constraints.
It would be better if this algorithm can transform these nonlinear constraints into the equivalent linear ones.
How can implementation an Adaptive Dictionary Reconstruction for Compressed Sensing of ECG Signals and how can an analysis of the overall power consumption of the proposed ECG compression framework as would be used in WBAN.
To solve a highly nonlinear function. We can not show analytically this function is quasi-concave.
NMPC- Nonlinear model predictive control
X_dot=f(x,u)
Y=C*x
objective function: min J = (Y-Ys)^2+du^2+u^2
w.r.t u
constraints are :
0<u<10
-0.2<du<0.2
0<Y<8 here Y is a nonlinear constraint and a vector
Dear Friends and colleagues
I have an optimization in which I have a nonlinear term in the following form:
x(t)* a(k)
where, x and a are variables. a is a binery variable and the sets in which each of the variables are defined is not the same. Would you please suggest me a method that I can use to handle this term and transfer my model to a mixed integer linear programming?
Thank you for your suggestions.
best (achievable) x-reg for singular solutions of nd incompressible euler equations by Székelyhidi jr.+de lellis and remarks/links with pressureless (regular) sols ?
--on solutions "convex optimized" by l. Székelyhidi jr. and c de lellis (and descendants) of nd (or 2d, 3d if there is a difference for LS-CD frames and if so, then how ?) incompressible euler equations, could anyone tell what is the best x-reg (so better than C ^ s, s about 1/3) that LSjr+CD or their descendants, 1 / have already, 2 / could be expected and what are the obstacles getting (or not) a better x-reg ? (all with the same question in t-reg and mixed (x, t) -regs), this regardless of their first motivations, that is, breaking the uniqueness (more generally realistic up to C ^ s for all s <1) and onsager (related to s=1/3) or the final answer is close to s=1/3 (and then why)?
--Can anyone confirm that the solutions coming from the classical and regular theory (that is to say not LS&CD theory), which are the particular solutions that are pressureless (see in 2 / 3d, majda diperna pl lions but also in Rn, all n, in my jmpa95 and thesis92-ch3/90 etc, quoted by yudovich) are out of reach by LSjr+CD theory and their singular frameworks (because for LSjr+CD, the pressure is, at first, basically + - equal to cst lul ^ 2 and then p = 0 gives u = 0 or all of this can be (partly) overcome and how, for example, include these sols in LS&CD setting?) = What could be said about these sols regarding the LS&CD theory ?
--regular theory =eg, for a (short) interval of non nul times, +- Du or rotu (:Du-tDu) are in x, C^s (s in ]0,1[), C^o, or L°° or bmo or (eg in 2d:) Du or rotu in Lp(loc), for one p in [1,oo] etc
--LS&CD singular theory at the state (?)= eg , for a (short) interval of non nul times, u itself (and not its gradient Du) is only in C^s, s in ]0,1[ or s<1/3 or s= 1/3 etc and eg nothing (yet?) on (atleast L1(loc)/loc-measures) DERIVATIVES of u, eg Du or rot(u) obtained (and conserved) for non nul times (even if it is supposed at t=0) etc
-- Are there (zones of) junctions or intersections (and where and what) between regular theory and LS&CD singular theory, existing same common sols (even particular) for these 2 theories ?
--eg 2d, or 3daxi, rotu (and/or its moments 0,1,2 and/or the axi structure) could be? (even "half" or "under" or partly in some way) conserved in SDth as in regTh etc ?.
--questions extended of course to all flu mech models already treated (by extensions from IncEE) by SDth. 6/7/18
what is the best method for constrained optimization ( in terms of the speed of the convergence an accuracy)?
I want to train the parameters of the neural network with some constraints on the output of the network, so I need an constrained optimization method. the number of parameters ( weights and biases of neural network) is about 200 to 1000 parameters.
I am reading Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers by Boyd (https://web.stanford.edu/~boyd/papers/pdf/admm_distr_stats.pdf), where the authors showed that ADMM converges under 2 assumptions (page 16):
1. the objective function f and g are closed, proper and convex.
2. the unaugmented Lagrangian $L_0$ has a saddle point.
A proof is given in Appendix A in the above PDF (page 106-110).
I am working on a problem which I would like to solve using ADMM, but the objective functions in my problem are not closed (but they are differentiable and convex). In this case, can I still use ADMM and have guaranteed convergence?
I am asking this because from the proof mentioned above, it seems that the assumption 1 is only used to justify the sub-differentiability of $L_\rho$, so I think I can still use ADMM in my problem given that $L_\rho$ is differentiable.
Would you please explain to me what is the technical definition or explanation behind the point that, in a convex optimization program or space, the equality constraint must be affine.
Hi,
I am working with a VERY large scale LP -- so large that simplex method takes forever to run. I've developed an efficient numerical algorithm to exploit the problem structure to significantly reduce the running time.
The problem is, my application requires basic solutions. For now, I am using crossover, that is, take the optimal primal and dual solutions (numerical) from my algorithm, give them to a simplex solver, and use good old simplex to solve it to get basic solutions. This works well, but it's still a bit too slow, as most of the running time are used in the crossover phase.
So my question is, is there any other algorithm in the literature that can produce basic solution given a numerical solution with high accuracy?
Thanks,
Alex
I am trying to solve SDPs using mosek where IPM is used. For very large SDP problem mosek has some issues. Is there any other solver available which uses different algorithm and can solve large scale SDPs ??
In solving sparse representation problems in image processing generally the L1-minimization techniques gives good result compared to other greedy algorithms. But sparse coding method that solves the minimization based on 'Feature-sign search algorithm' is slow. Does the methods like, iterative shrinkage thresholding algorithm (ISTA) or FISTA (fast ista) which are commonly used in image denoising problems, will be more efficient if it is used to replace the 'Feature-sign search algorithm' in image super-resolution ?
links:
How do we carry iteration of matrix rank approximation and what is the initial value and step size? M is matrix with missing values. We approximate X to M. Its in matrix completion problem. Any answer is appreciable.
I have studied and understood the Moment-SOS hierarchy proposed by Lasserre where a sequence of semi-definite programs are required to be solved and a rank condition is invoked in order to check if the global solution is found. I was not able to find such conditions for its dual viewpoint ( also known as the Putinar's Positivstellensatz). Alternatively, is there a similar rank condition for Parrilo's sum-of-squares relaxation?
I want to know what will be the dual of the problem in attached file.

I am new to this field of optimization, if someone could suggest some materials to study convex optimization, that would be better. I have started reading S Boyd. But the geometric interpretation of all the things like convex function, affine function, cone, convex hull, conic hull etc are not very clear with that.
I have to design a controller using Convex Optimization and Sum of Squares technique, please suggest some papers on this. What are the advantages of using this techniques from the classical controllers or other robust controllers.
does it mean that the two functions are not antagonists? I have solved the optimization problem using the weighted sum method in MATLAB .
thank you

Solving large multi-stage stochastic problems may become intractable. There are some applications permit to solve such problems, among which "non-anticaptivity principle", "decomposition techniques", "lagrangian relaxation", and lately "optimal condition decomposition" are the most known ones. In this regard, what are the advantages and disadvantages of the "scenario reduction techniques" attribute to mentioned techniques to reduce computational burden as well as complexity with a good approximation to the original stochastic optimization problem? In advance, I'd appreciate your supportive message.
Is there a definition of convex optimization problem for COMLEX-VALUED MATRIX VARIABLES where objective and constraint functions are real-valued ?
Are KKT conditions true for convex optimization problem with COMPLEX-VALUED MATRIX VARIABLES where objective and constraint functions are real-valued ?
They are often seen in MIMO communication problems.
Thank you very much.
I have a system of n atoms and I am using a nonlinear conjugate gradient (Polak–Ribière) to minimize its energy, which is the Lennard Jones potential. I use strong Wolfe conditions for the line search. The code works great and fast for small systems (up to 60 atoms). But for larger systems the method gets very slow due to round off errors. The step size found using strong Wolfe conditions is small and so for large systems each atom must move a very short distance (because this small step is multiplied by atoms positions then added to previous position vector for the whole system) and on a computer the system does not feel it.
I want to maximize a unknown N-dimensional function f(x1,x2,...xN)
each xi is independent each other, xi>0 for all xi, and f is smooth.
My function is not expensive, so I tried f(x1+dx1,x2+dx2,...) with randomly generated dx1,dx2...
However, I have to calculate f more than millions of times, and sometimes f stuck in local maxima and doesn't give maximum value.
Is there any faster and simple algorithm that can maximize f?
Thank you in advance.
Specifically, estimation of parameters in complex probability distributions
- Variational Inference.
- MCMC
Hi,
I am trying to use CVX to implementing the antenna selection problem proposed in the following paper:
A. Dua, K. Medepalli e A. Paulraj, "Receive antenna selection in MIMO systems using convex optimization", IEEE Transactions on Wireless Communications, vol. 5, n.o 9, pp. 2353-2357, 2006.
The problem consists of a channel maximization problem, described as follows:
cvx_begin
variable x(Mr)
maximize( log_det( eye(Mr) + EbN0*diag(x)*Hfull*Hfull' )/(log(2)) )
subject to
0 <= x <= 1;
trace(diag(x)) == L;
cvx_end
where Mr is the number of antennas in the full set, Hfull is the full channel matrix and L is the number of antennas to be selected.
The problem is that the CVX is returning 'x' with all elements equal to each other.
Does anyone know why this is happening and how I can solve it?
PS: This is a linear relaxation of the optimum problem, and I am supposed to get the L larger entries of 'x' as the selected antennas.
Thank you.
Regards.
hi dear all; please consider a geometric programming problem such as attached example. this is the cost function of a GP problem. can this function be a convex function without any variable change?

hi dear all; consider 5 complex nonlinear equations with 5 variables. what is the simplest method for solution?
I have used epsilon constraint method as well ass the sum weighted method to find Pareto point for bi-objective model;in this case I have found same results ,same number of Pareto points,
can we claim which our bi-objective model is not non-convex if the results raised from epsilon constraint method and the sum weight method are same?
I need to find a entropy minimizing probability distribution under a convex constraint.
The solution of bi-level model is too hard. Generally, duality, decomposition and evolutionary algorithms are used for solving.
i have convex problem that start solved for first 50 iteration the after that its go to inaccurate solved .How to resolve the problem?
Is there any condition (similar to K.K.T condition in convex optimization) can be used to get the analytical solution.
I want to get some theorems form the optimization model, and analyze the relationship between the variables and the parameters.
How about the Multi-modules (used to prove the optimality of the model) ?
One optimization problem is called system-centric optimization, which is the objective of the authority.
Another set of optimization problems is called user-centric optimizations, which represents the objectives of the users/ the customers.
For the authority, they of course hope to achieve the system-centric goal. So, the actual final result is based on user-centric optimization. However, the users/customers only hope to maximize their own utility which is somehow contradicted with the system-centric optimization.
However, the authority can adjust his one or more parameters in user-centric optimization problems. How to adjust those parameters so the user-centric optimization's result/solution is aligned with that of the system-centric optimization? Or how to adjust those parameters so that the objective function of the system-centric optimization problem can be minimized (Suppose it's a minimization problem).
Nesterov proposed accelerated gradient methods for optimization. Have other such approaches been proposed.
What is use of xmax in Quantum integrated Particle Swarm Optimization (QPSO)? How can I define xmax in QPSO? How it signifies QPSO?
please consider maximum allowable power PT in an OFDM power allocation problem when Pi is the power value of each subcarrie . can we use equality constraint: SUM(Pi)=PT instead of SUM(Pi)<PT inequality?

i was wondering if there is any approach that can reformulate clustering algorithm by standard optimization and maybe relax it into a standard convex optimization with generalized inequality in proper cones like SDP or SOCP.?
I need to theoretically prove the convergence rate of my optimization algorithm. Please suggest some review papers or materials to understand the concept of convergence rate, its different types and process of derivation of convergence rate.
I know that there are elegant formulas for the KKT conditions of conic optimization problems with convex objective function for particular cones (such as for example the direct product of Lorentz cones). Are there similar formulas for a general cone?
How to find the trade-off point when both objective functions in a multi-objective optimization problem are subjected to maximization, while the obtained Pareto front become concave up, decreasing (like what is common in minimization problems) ?

Can you provide me with such an example (even in the finite dimensional setting) ?
There are forward-backward, Douglas-Rachford, and primal-dual schemes in the splitting methods family. But in the literature, it is very rare to see the derivation of these methods. The authors usually give fancy and sometimes elegant iterative expressions followed by the proof that the proposed method converges.
It seems that the root of splitting methods are from the subgradient of f(x)+g(x) contains zero. But it also seems that there are many ways after that to get a fixed point. In addition, there are more general splitting methods that handle the summation of three terms or more.
Could anyone give some suggestions and point out a reference? Thanks.
Hello;
I have ended up with a quadratic function of X in the form of:
f(X)= XTAX-XTBX+CTX with A and B as positive definite matrices. Note that the second term has a negative sign. Clearly the function is not necessarily convex, but based on some experimentation and prior calculations, it should convex. I wonder to know if there is any analytical or famous experimental method to prove the convexity of this quadratic function.
Thank you
I have formulated optimization problem for building, where cost concerns with energy consumption and constraints are related to hardware limits and model of building. To solve this formulation, I need to know if problem is convex or non-convex, to select appropriate tool to solve the same.
Thanks a million in advance.
I am looking for convex optimization solver based on Hadoop MapReduce. I am also interested to convex optimization in a Big Data context.
Best regards,
Sabeur
I am working on a part of a paper related to topological properties of boundary points. It is important for me to realize the topological and algebraic behavior the boundary points of the convex sets. I would be grateful if someone could help me around this issues by giving some ideas or references related to it.
The general question provided in following;
Let B be a closed set in n−dimensional Euclidean space. What other properties B should have in order to be guaranteed that there exist the closed convex set A such that ∂A=B. How about infinite dimensional spaces?
The problem is: min f(x),
s.t: Ax>=b,
where f is convex.
In practice, how does one normally detect if the problem is in-feasible when optimizing using ADMM methods ?
I need to comment whether my optimization function is convex or non-convex. My optimization function is in the form of (y-y_cap)^2. y is know. y_cap comes out of a MATLAB pfile. So, y_cap can be considered output of a black-box model. It is also not possible to plot y-y_cap as it is a 22*8 matrix
In convex optimization, what is the definition of the facet of a convex hull, and how to prove a constraint is facet-defining? Can anyone recommend some related references?
For a given pdf function f(x) defined in [0,1], which is monotonically increasing and convex, I need to find a probability generation (polynomial) function Q(x) of a maximum degree D, where its derivative is always larger than or equal to the pdf function in [0,1-e], for every e>0, Q'(x)>f(x). It can be solved by numerical optimization algorithms, but I am looking for some theorems to prove the existence of such a PGF for every e.
Please find here more details on the question:
I want to use the SOSTOOLS toolbox, which works in Matlab and can be combined with the following SDP solvers: CSDP, SDPNAL, SDPA, SeDuMi, and SDPT3. I'm not sure about which of these solvers I should choose. Any advice would be welcome.
I would really appreciate if anyone can give me a formal citation for this. I know this is a standard problem for which I also have a proof. However this is requried as a part of a more complicated problem and it would be nicer for the writing if I can just simply give a citation.
There are efficient methods in DC programming. On the other hand every C1 function with a Lipschitz derivative on a compact convex set is the difference of a convex function and a convex quadratic function, e.g., J. Glob. Opt. 46(2010)155-161. This means that one can optimize C1 functions by DC methods. Has anybody tried this approach?
As a part of my work I need to obtain all the extreme optimal solutions of a given LP. So far I came up with the solution where I find the value of the LP, fix it as a constraint and find all the feasible extreme solutions of the resulting LP, which leads to finding all the vertices of the polyhedron formed by the constraints of this LP. Can somebody please direct my to some work where there is a worst case complexity analysis of finding all vertices of such polyhedron?
ABSOLUTE VALUE OPERATOR LINEARIZATION
I have a nonlinear term in the objective function of my optimization problem as an Absolute Value function like |x-a|.
As far as I know, an Absolute Value operator makes the optimization problems nonlinear (i.e. NLP). How can I make it linear (LP or MILP)?
Max f(x)=g(x) + b*|x-a|
s.t. some linear constraints
Regards,
Morteza Shabanzadeh
Binary Variable * Real Variable = ?
1) lead to an equivalent 'Nonlinear' variable (and thus => MINLP),
2) lead to an equivalent 'Integer' variable, 'Discrete' I mean (and thus => MILP).
Which one is correct and why?
What is your idea to deal with this problem by adding a constraint and make the resultant problem MILP (if it is not MILP).
Regards,
Morteza Shabanzadeh
Since the strict complementarity condition (between the Lagrange multipliers and the inequality constraints) is not guaranteed for the optimal solution of a quadratic programming problem, I wonder if how I can have the closest solution to the optimal one which holds the strict complementarity condition.
I have an optimisation problem with a two-stage cost function, consisting of two parts. For a parameter value less than a threshold (which is an optimal value by itself), only the first part appears in the optimal value function. For the parameter value more than this threshold, on the other hand, both parts appear in the optimal value function.
I tried to under-estimate the problem by under-estimating all non-convex terms in the cost function and constraints; however, the optimal value function of the resulting under-estimated problem is not convex because of the two-stage operation.
I am wondering if there is any way to under-estimate such a problem considering the multi-stage operation in order to have a convex optimal value function of the resulting under-estimated problem.
How do you minimize the following non linear optimization problem:
find x* that minimizes the H-infinity-norm: || G(s, x) ||oo
Subject to : xmin < x < xmax
Where:
s=j*w and w = wmin to wmax { for example: w = logspace (-3,+3,50)}
G: denotes the transfer matrix that given (for example as):
G(s,x)=[(s^x1)+3/s -9+x2*s+x3; 1+x1^x2 s/x3]
x=[x1;x2;x3] denotes the design parameter vector to be determined by an optimization method
in convex optimization who can explain what wi will say by this :
minimize 1/2 xTAx+cTx
subject to Dx=b
0<=x<=d
?
Software to linear programming as cplex, gurobi, etc but with the method of aggregation of columns integrated(Dantzing wolfe)
Is possible save memory in big problems with this method?
Hello,
For those of you who are familiar with Pincus Theorem which is based on finding a global maximum of ANY objective function.
Refer to page 63 here, http://www.iis.sinica.edu.tw/page/jise/2001/200101_04.pdf
Can anyone, kindly, provide me with another similar theorem which finds the global maximum of any objective function (No constraints here).
Consider the following setting: g(y,V) = minw\in C(y-w)TV(y-w), C is a closed convex cone, V is positive definite and y is in Rn.
Assume an appropriate norm for (y,V), would g(y,V) be continuous in (y,V) ?
I know if C is compact then this is easy to prove. But a cone cannot be compact.
At first glance I thought this should be a trivial problem but whatever strategy I tried I got stuck in the compactness of C. So I turn to the literature and it turns out so far that continuity of such a function like g(y,V) seems to be an active area of research ( I do hope I am wrong!)
I am very naiive to convex optimization in general, particularly conic projection. So I will appreciate any hints or comments on how to tackle this problem as well as solid disproof!
Many thanks in advance
For g(y,V,w)=(y−w)TV(y−w), for y in Rn, V is positive definite (e.g. variance matrix) and w is in a closed convex cone C in Rn.
Obviously for every fixed y and V minw\in C g(y,V,w) exists
But my question is for every fixed y,V and y0, V0 does
minw\in C [ g(y,V,w) - g(y0,V0,w) ] exist?
Hello,
I would like to know that how I can find the number of variables (especially the integer ones) in GAMS (General Algebraic Modeling System) codes.
Does GAMS platform have any options to show the number of variables?
Any help would be appreciated.
Regards,
Morteza
For example, if I have in matlab matrix of form [A*B;A*C;A*D] how I can check if it represents the result of Kronecker product?
In a Euclidean space, an object S is convex, provided the line segment connecting each pair of points in S is also within S. Examples of convex objects in the attached image include convex polyhedra and tilings containing convex polygons. Can other tilings containing convex shapes be found?
Solid cubes (not hollow cubes or cubes with dents in them) are also examples of convex objects. However, crescent shapes (a partial point-filled circular disk) are non-convex . To test the non-convexity of a crescent, select a pair of points along the inner edge of a crescent and draw a line segment between the selected points. Except for the end points, the remaining points in the line segment will not be within the crescent. Except for the 3rd and 5th cubes, the cubes in the attached images are convex objects (all points bounded by walls of each cube are contained in the cube).
From left-to-right, the cresent shapes are shown in the attached image are non-convex: Nakhchivan, Azerbaijan dome, Taj Mahal, flags of Algeria, Tunisia, Turkey and Turkmenistan. For more examples of crescent objects, see
Can you identify other crescent shapes in art or in architecture that are non-convex? Going further, can you identify other non-convex objects in art or in architecture?
The notion of convexity leads to many practical applications such as optimization
image processing
and antismatroids, useful in discrete event simulation, AI planning, and feasible states of learners:
In science, convex sets provide a basis solving optimization and duality problems, e.g.,
Convex sets also appear in solving force closure in robotic grasping, e.g.,
Recent work has been done on decomposing 2D and 3D models into their approximate convex components. See, for example, the attached decompositions from page 6 in
J.-M. Lien, Approximate convex decomposition and its applications, Ph.D. thesis, Texas A&M University, 2006:
There are many other applications of the notion of convexity in Science. Can you suggest any?
Conference Paper Projection on Convex Set and Its Application in Testing Forc...



