Science topic
Applied Optimization - Science topic
Explore the latest questions and answers in Applied Optimization, and find Applied Optimization experts.
Questions related to Applied Optimization
Hello fellow researchers,
I'm writing to suggest a mutual citation exchange to encourage collaboration and support within our academic community. I have recently published the following papers and would greatly appreciate citations from fellow researchers in related fields.
In return, I am more than willing to reciprocate by citing your work in my future publications. Below I have provided a list of articles for your consideration:
Camargo, F. G. (2021b). Survey and calculation of the energy potential and solar, wind and biomass EROI: application to a case study in Argentina. DYNA, 88(219), 50-58. https://doi.org/10.15446/dyna.v88n219.95569
Camargo, F. G. (2022c). Dynamic Modeling Of The Energy Returned On Invested. DYNA, 89(221), 50–59. https://doi.org/10.15446/dyna.v89n221.97965
Camargo, F. G. (2022d). Fuzzy multi-objective optimization of the energy transition towards renewable energies with a mixed methodology. Production, 32, e20210132. https://doi.org/10.1590/0103-6513.20210132
Camargo, F. G. (2023e). A hybrid novel method to economically evaluate the carbon dioxide emissions in the productive chain of Argentina. Production, 33. http://dx.doi.org/10.1590/0103-6513.20220053
Camargo, F. G., Schweickardt, G. A., & Casanova, C. A. (2018). Maps of Intrinsic Cost (IC) in reliability problems of medium voltage power distribution systems through a Fuzzy multi-objective model. Dyna, 85(204), 334-343. https://doi.org/10.15446/dyna.v85n204.65836
Please feel free to reach out if you're interested in this collaboration or have any questions. Looking forward to connecting and exchanging citations!
Best regards,
PhD Camargo Federico Gabriel
Technology Activities and Renewable Energies Group
La Rioja Regional Faculty of the National Technological University, Argentina.
If you want to do any collaborations, feel free to contact with my team.
latine hyper cube design
global and local optimization
4D color maps How do we assess the global behavior of a model and obtain the equilibrium points and analyze their stability through simulation series ?
No one has the mental capacity to know all languages. Additionally, the more languages one is fluent in, the more likely that individual will mix up words. Thus, knowing enough languages for survival is optimal while artificial intelligence could and potentially will bridge language barriers. Of course knowing three languages or more is somewhat of an advantage.
The set of optimal solutions obtained in the form of Pareto front includes all equally good trade-off solutions. But I was wondering, whether these solutions are global optima or local optima or mix of both. In other words, does an evolutionary algorithm like NSGA-II guaranties global optimum solutions?
Thank you in anticipation.
Hello!
Multiple Criteria Decision-Making (MCDM) methods are applied in many fields of science, as a result, many scientific publications related to the application of these methods have been prepared.
Some of the most popular MCDM methods or MADM (Multiple Attribute Decision-Making) are TOPSIS, SAW, AHP, etc. In describing these methods, some authors use the term "criteria", and others use the term "attribute". I would like to know your opinion on which term should be used.
Some references:
Yoon, K. P., & Hwang, C. L. (1995). Multiple attribute decision making: an introduction. Sage publications.
Triantaphyllou, E. (2000). Introduction to Multi-Criteria Decision Making. In: Multi-criteria Decision Making Methods: A Comparative Study. Applied Optimization, vol 44. Springer, Boston, MA. https://doi.org/10.1007/978-1-4757-3157-6_1
Thank you!
Over the last few decades, there have been numerous metaheuristic optimization algorithms developed with varying inspiration sources. However, most of these metaheuristics have one or more weaknesses that affect their performances, for example:
- Trapped in a local optimum and are not able to escape.
- No trade-off between the exploration and exploitation potentials
- Poor exploitation.
- Poor exploration.
- Premature convergence.
- Slow convergence rate
- Computationally demanding
- Highly sensitive to the choice of control parameters
Metaheuristics are frequently improved by adding efficient mechanisms aimed at increasing their performance like opposition-based learning, chaotic function, etc. What are the best efficient mechanisms you suggest?
A collection of solved examples in Pyomo environment (Python package)
The solved problems are mainly related to supply chain management and power systems.
Feel free to follow / branch / contribute
Hello,
I am trying to run an optimal power flow (OPF) study with Matpower. In the standard OPF formulation, only the P and Q generation of generations are part of the objective function (or cost function to be minimized).
However, I would like to have the bus voltages in the objective function, so that for example the total voltage deviation of all voltages can be minimized. Does anyone have experience with this?
Thanks.
Could any expert try to examine our novel approach for multi-objective optimization?
The brand new approch was entitled "Probability - based multi - objective optimization for material selection", and published by Springer available at https://link.springer.com/book/9789811933509,
DOI: 10.1007/978-981-19-3351-6.
I am trying to convert vector into an image using the code below
clear variables
load(Exe4_2022.mat')
n = length(b);
figure,
imagesc(reshape(b,sqrt(n),sqrt(n))),
colormap(gray),
axis off;
But I am getting this error. Could anybody tells me how to resolve this issue??
Error using reshape
Size arguments must be real integers.
I have attached the "Exe4_2022.mat" file with this post.
Thanks
This is in context of the objective function of a multivariate optimization problem say, f(a,b,c).
I am looking for a "measure" for the degree of bias of f(a,b,c) towards any of the input variables.
What is the main disadvantage of a global optimization algorithm for the Backpropagation Process?
Under what conditions can we still use a local optimization algorithm for the Backpropagation Process?
I have 2 functions f(x,y) and g(x,y) that depend on two variables (x,y), so I want to find a solution that minimize f(x,y) while maximizing g(x,y), simultaneously??
P.S: These functions are linearly independent.
Can anyone provide me with PSO MATLAB code to optimize the weights of multi types of Neural Networks?
I would like to optimise the process model of a thermal energy supply system, which was developed in the software environment IPSEpro, with regard to the economic and energetic constraints. Since the options in the software itself are limited in terms of optimisation algorithms, I would like to optimise the process model via the COM interface with Matlab with the help of the optimisation algorithms available in Matlab. What is the best way to link the Matlab code for optimisation with the code that controls the call to the external process model? How can the objective function for the algorithm be formulated in Matlab when there is no functional relationship, but a parameter of the external model should be used?
For an Integer Linear Programming problem (ILP), an irreducible infeasible set (IIS) is an infeasible subset of constraints, variable bounds, and integer restrictions that becomes feasible if any single constraint, variable bound, or integer restriction is removed. It is possible to have more than one IIS in an infeasible ILP.
Is it possible to identify all possible Irreducible Infeasible Sets (IIS) for an infeasible Integer Linear Programming problem (ILP)?
Ideally, I aim to find the MIN IIS COVER, which is the smallest cardinality subset of constraints to remove such that at least one constraint is removed from every IIS.
Thanks for your time and consideration.
Regards
Ramy
I am coding a multi-objective genetic algorithm, it can predict the pareto fronts accurately for convex pareto front of multi-objective functions. But, for non-convex pareto fronts, it is not accurate and the predicted pareto points are clustered on the ends of the pareto front obtained from MATLAB genetic algorithm. can anybody provide some techniques to solve this problem. Thanks in advance.
The attached pdf file shows the results from different problems
I am solving Bi-objective integer programming problem using this scalarization function ( F1+ epslon F2). I have gotten all my result correct but it says Cplex can not give an accurate result with this objective function. It says cplex may give approximate non-dominated solution not exact. As I said before, I am very sure that my result is right because I already checked them. Do I need to prove that cplex give right result in my algorithm even sometimes it did mistake in large instance?
Thanks in advance.
Hi all,
I have a large mixed-integer programming (MIP) optimization problem, which has a high risk of infeasibility. The branch and cut algorithm in GLPK spends hours to find an optimum solution, and it may return an infeasible solution at the end. I want to do a pre-screening before starting the actual optimization to make sure there is a good chance of a feasible solution. I admire the fact that the only way to check feasibility is to run the optimization, but any heuristic with potential false infeasible alerts (false positives) could be helpful. My focus is on feasibility rather than optimality. Do you have any suggestion of algorithm, software, or a library in Python to do this pre-screening?
Thanks for your time and kind reply.
I optimized 02 structures containing NO2 group by Gaussian at 6 311G (d,p) level. In the output file i observed that the NO2 is not connected to the structure and appears as O=N=O .
Q1: When i to use this output as a starting structure for a TS search; should i reconnect the NO2 group to the structure by single bond or i have to keep the output structure as it is ?
i tried so many Keywords such as opt=(calcfc,ts,noeigen), opt=(calcall,ts,noeigen), # opt=(calcfc,tight,ts,noeigentest)... and many guesses but the irc showed that the TSs is not connecting the reagents and products!
Q2: is there any other option tofind the right TS of this pathway!
Any help is much appreciated
Hi
I have a project in the field of optimizing the groundwater monitoring network through coding in MATLAB software (with NSGA2 algorithm), I have read the complete research background and I am completely familiar with the subject in theory, but I have no background to start coding. Does anyone have a related or educational code file in this subject?
Thank you for your help
Dear all, I want to solve three or four objective functions with the four or five decision variables (no. of decision variable> no of the objective function). There may be some game-theoretic approach to solve the problem. But I am searching for some MATHEMATICA code (Like the Direct Search method or other computational technique) to solve the problem. If there have, please suggest (other than GA).
Thank you.
Hello scientific community
Do you noting the following:
[I note that when a new algorithms has been proposed, most of the researchers walk quickly to improve it and apply for solving the same and other problems. I ask now, so why the original algorithm if it suffer from weakness, why the need for a new algorithm if there are an existing one that solved the same problems, I understand if the new algorithm solved the unsolved problem so welcome, else why?]
Therefore, I ask, is the scientific community need a novel metaheuristic algorithms (MHs) rather than the existing.
I think, we need to organized the existing metaheuristic algorithms and mentioned the pros and cons for each one, the solved problems by each one.
The repeated algorithms must be disappear and the complex also.
The dependent algorithms must be disappeared.
We need to benchmark the MHs similar as the benchmark test suite.
Also, we need to determine the unsolved problems and if you would like to propose a novel algorithm so try to solve the unsolved problem else stop please.
Thanks and I wait for the reputable discussion
Dear all,
I want to start learning discrete choice-based optimization so that I can use it later for my research works. I want to know about free courses, books, study materials available on this topic. Any suggestions will be appreciated.
Thanks,
Soumen Atta
For example, I have known variables x and y have nonlinear relationships with the score, however the model of x, y to the score is unknown. We want to know which values of x and y can lead to the highest possible score y. Could you kindly recommend related methods or researchs?
Hi,
I'm interested in solving a nonconvex optimization problem that contains continuous variables and categorical variables (e.g. materials) available from a catalog.
What are the classical approaches? I've read about:
- metaheuristics: random trial and error ;
- dimensionality reduction: https://www.researchgate.net/publication/322292981 ;
- branch and bound: https://www.researchgate.net/publication/321589074.
Are you aware of other systematic approaches?
Thank you,
Charlie
I am working on a problem for design optimisation. I would like to ask if for an uncertain problem, should design optimisation under uncertainty techniques be used for the design optimisation?
Hi,
I'm running BARON on an AMPL instance that I uploaded on the NEOS server. Unfortunately, it times out after a few minutes. It is possible to pass an option file to set the time limit (maxtime), but I struggle to find the right syntax.
I've tried:
option maxtime 10000
maxtime 10000
MAXTIME=10000
Either it is not recognized, or it has no effect.
Can you help me out?
Thanks,
Charlie
I am using stochastic dynamic dual programming for decision-making under uncertainty. Can we use stochastic dynamic programming to solve a min-max problem? for example, the max function is used in the objective function. Any good library for stochastic dynamic dual programming?
I am using the ANN for a product reliability assurance application, i.e.picking some sample within the production process and then estimating the overall quality of the production line output. What kind of optimization algorithm do you think works the best for solving the ANN in such a problem. ?
I am working on ECG arrhythmia classification by using SVM , implemented some kernels tricks
and using different kernels on MIT BIH dataset (features create 44187 row ,18 column matrix)
now it is difficult to plot support vector for such large data sets , now how can i plot it and please suggest any other plots or methods to show comparison between different kernels , i already have comparison chart of accuracy efficiency etc.
Any decision-making problem when precisely formulated within the framework of mathematics is posed as an optimization problem. There are so many ways, in fact, I think infinitely many ways one can partition the set of all possible optimization problems into classes of problems.
1. I often hear people label meta-heuristic and heuristic algorithms as general algorithms (I understand what they mean) but I'm thinking about some things, can we apply these algorithms to any arbitrary optimization problems from any class or more precisely can we adjust/re-model any optimization problem in a way that permits us to attack those problems by the algorithms in question?
2. Then I thought well if we assumed that the answer to 1 is yes then by extending the argument I think also we can re-formulate any given problem to be attacked by any algorithm we desire (of-course with a cost) then it is just a useless tautology.
I'm looking foe different insights :)
Thanks.
I want to use optimization in classification tree tasks. I am not sure how can I do that?
Bat-inspired algorithm is a metaheuristic optimization algorithm developed by Xin-She Yang in 2010. This bat algorithm is based on the echolocation behaviour of microbats with varying pulse rates of emission and loudness.
The idealization of the echolocation of microbats can be summarized as follows: Each virtual bat flies randomly with a velocity vi at position (solution) xi with a varying frequency or wavelength and loudness Ai. As it searches and finds its prey, it changes frequency, loudness and pulse emission rate r. Search is intensified by a local random walk. Selection of the best continues until certain stop criteria are met. This essentially uses a frequency-tuning technique to control the dynamic behaviour of a swarm of bats, and the balance between exploration and exploitation can be controlled by tuning algorithm-dependent parameters in bat algorithm. (Wikipedia)
What are the applications of bat algorithm? Any good optimization papers using bat algorithm? Your views are welcome! - Sundar
Hello everyone,
We have the following integer programming problem with two integer decision variables, namely x and y:
Min F(f(x), g(y))
subject to the constraints
x <= xb,
y <= yb,
x, y non-negative integers.
Here, the objective function F is a function of f(x) and g(y). Both the functions f and g can be computed in linear time. Moreover, the function F can be calculated in linear time. Here, xb and yb are the upper bounds of the decision variables x and y, respectively.
How do we solve this kind of problem efficiently? We are not looking for any metaheuristic approaches.
I appreciate any help you can provide. Particularly, it would be helpful for us if you can provide any materials related to this type of problem.
Regards,
Soumen Atta
Is there a Python project where a commercial FEA (finite element analysis) package is used to generate input data for a freely available optimizer, such as scipy.optimize, pymoo, pyopt, pyoptsparse?
There are many research on metaheuristic optimization e.g. Particle Swarm Optimization, Genetic Algorithm, etc. Some study show that they are good for clustering task. But I don't find any comparation about it.
Which one is the best to be applied for optimizing the clustering process?
Hi,
I'm a researcher in optimization and a hobbyist photographer, and I'd like to get acquainted with lens design via the use of optimization methods. I found for example the paper "Human-competitive lens system design with evolution strategies" (2007).
Are you aware of more recent techniques to design camera lenses? Are there optimization models or benchmarks available?
Thank you,
Charlie
What are some of the well-written good references that discusses why finding the penalty parameter when solving a nonlinear constrained optimization problem is hard to find from the computational perspective. What are some of the computational methods done to find the parameter as I understand finding such a parameter is problem-dependant.
Any insights also would be very helpful.
Thanks
Hello everyone,
My issue is about a water distribution system that I am working on a zone of the system where does not exist any plan for the place of pipes. However, we have the place of actuators like different types of valves, pressure relief valves, pressure meters, flow meters and tanks. Also, we have the place of demands where suffer from pressure loss. Now the question is how we can do pressure management using actuators to maximize water pressure for all demands of the zone based on the previous recorded data while we are going to minimize water loss as well as pipes damaging. Please let me know if you have any idea or you know any suitable paper for this issue.
Thanks.
Dear all
I am working on an inventory model in closed-loop supply chain system to optimize the cost of the system. There are lots of model to optimize the cost of the system, but I am looking forward to incorporate the concept of the Analytics to handle the real time inventory.
Looking forward to hearing from you.
with regards
Sumit Maheshwari
Dear esteem researchers,
Greetings!
I applied a Markowitz risk theory on my power system operation objective function. In order to calculate the risk cost, i need to obtain the variance of the objective function which result to non-linear (the objective function is raised to power 2).
i know i can linearise the problem using piecewise linearization approach, however, my challenge is how to determine the segment/ grid points interval of the problem since by objective function is embedded with some decision variables with different bounds.
Please, your support and any recommendation will be highly appreciated.
Thanks
Four years ago I was working on a genetic algorithm for vectorization and colors reduction. It is an open source project available in GitHub under the name EllipsesImageApproximator. The result of the algorithm is a list of simple shapes (in this case ellipses). Now I need this list to be translated in G Code instructions. I am working with 16 base colors. Each color will be plotted separately by the plotter (color by color plotting). For 16 colors there will be 16 CNC files with G Code instructions.
Please, suggest me what will be the most efficient way for G Code instructions generation?
I am working on the optimization problem of multi-energy system using benders decomposition algorithm approach. I have written the subproblem and the master problem separately as a function using Matlab, but the benders algorithm I developed is not working as expected. Please your support will be appreciated, I need benders algorithm template in Matlab.
Thanks,
Michael
We have a stochastic dynamic model: Xk+1 =f(Xk,uk,wk ). We can design a cost function to be optimized using dynamic programming algorithm. How do we design a cost function for this dynamic system to ensure stability?
In Chapter 4 of Ref. [a] for a quadratic cost function and a linear system (Xk+1 =AXk+Buk+wk), a proposition shows that under a few assumptions, the quadratic cost function results in a stable fixed state feedback. However, I think about how we can consider stability issue in the designation of the cost function as a whole when we are going to define the optimal control problem for a nonlinear system generally. Can we use the meaning of stability to design the cost function? Please share me your ideas.
[a] Bertsekas, Dimitri P., et al. Dynamic programming and optimal control. Vol. 1. No. 2. Belmont, MA: Athena scientific, 1995.
I would be grateful if anyone could tell me how the McCormick error can be reduced systematically. In fact, I would like to know how we can efficiently recognize and obtain a tighter relaxation for bi-linear terms when we use McCormick envelopes.
For instance, consider the simple optimization problem below. The results show a big McCormick error! Its MATLAB code is attached.
Min Z = x^2 - x
s.t.
-1 <= x <= 3
(optimal: x* = 0.5 , Z* = -0.25 ; McCormick: x*=2.6!)
I am working on a multi-criteria optimization problem, but I am facing problems to define proper fitness function.
Please, can you advise me on how you are choosing such fitness functions?
For a industrial application (layout-planning) I am currently trying to globally optimize a discontinuous function f. The objective function is defined on some bounded parameter space in R^N where the dimension N of this space depends on some initiation parameter. The N lies typically between 30-100.
The goal is to run this optimization a number of times (each time for a slightly different layout) and afterwards choose the best one.
Currently I use the MLSL-algorithm provided by the NLOPT-library to compute the global minimum of the objective. Especially when N goes up the time needed for each run to obtain a good result increases a lot. This is why I am looking for a way to speed up my computations.
From the structure of the objective function I know that it is oscillating which slows down the convergence of typical global optimization-algorithms. On the other hand the function f is the sum of a differentiable function and a upper-semi-continuous step-function, so in particular f is upper-semi-continuous and almost everywhere differentiable. The objective function is bounded as well and as it is defined on a bounded set it is integrable.
My question now is: Does anyone here have experiences with optimization of such functions (or more generally noisy or black-box functions) and has experiences which algorithms work best?
Especially as I have more details about my function is it maybe possible to use a subgradient-method or first smooth out my function by let's say a smoothing kernel phi, i.e. g = f * phi, and then optimize g to obtain a result for f?
I have programmed a method for solving quadratic optimization problems under linear constraints, this method is depends on the projection of a point onto a convex polyhedron in R^n, so I have programmed Dykstra's successive projection method and adapted it in my method, but Dykstra's successive projection algorithm dosen't work well, it's spend a lot of time even days to find a projection in 3 dimention real space!!!, I don't know if his algorithm is slow or I haven't programmed it properly!!! I have spend a lot of time on this method, so I'm very pluseur if somone can guide me to another projection method that I can get the algorithm code ready.
Hello ... the objective here is maximizing Z for each product (i)
function [ Zi ] = myfitness( X )
P=X(1);
C=X(2);
Q=X(3);
% Zi= fitness value
% C,P,Q = variables vectors
for i=1:10;
Zi = P(i).*Q(i)-C(i).*Q(i);
end
end
the outputs should be a 1*n matrice
when i run the function it works but i get only one value , and doesn't work with ga toolbox i keep getting the same error (index exceeds Matrix dimensions) ....how can i fix this error?
any help would be appreciated ....Thank you
Research pointed me to reinforcement learning. However, I will not be able to obtain a realistic simulation of the machine.
Currently I aquired about 5 months worth of data about target values, measurement values and quality of the produced parts as tabular data.
Therefore I had the following idea for a simulated environment:
- The RL agent chooses a set of target values based on the table.
- The agent receives a random observation (measurement values) that match the selected target values.
- The reward depends on the quality of the produced part that matches the selected target values.
--> The agent should then proceed to learn the optimal target values.
My questions:
- Are there better ways to simulate the environment?
- Are there better ways than reinforcement learning to determine the best target values?
Thank you for reading this!
Hello,
i'm looking for slsqp algorithm for optimization. I have to implement it on a non linear problem of minimisation with constraints in python and i'd like to know if it is appropriate for my problem.
Thanks in advance
Can higher order partial derivatives be derived or approximated from lower order partial derivatives?
There is no specific equation to state the partial derivatives, but they can be measured empirically.
Can higher order partial derivatives be derived from lower order partial derivatives, like 3rd (4th) order from 2nd (3rd) order? And how long can you continue this approximation of an order from the preceding order?
Empirically measuring the higher order partial derivatives is computationally too expensive in this case.
I have found that some mathematicians disagree with meta-heuristic and heuristic algorithms. However, from a pragmatic point of view such algorithms often can find high-quality solutions (better than traditional algorithms) when tackling an optimization problem but the success of such algorithms usually depend on tuning the algorithms' parameters. Why some mathematicians are against these algorithms? is it because they don't posses a convergence theory?
I am looking for different arguments on the situation.
Thank You!
- Any Scientific or empirical evidence / reasons why out of 196 countries in the world only 25 of them are very rich.
I am looking simple method to design 1*16 microstrip power divider for wideband (impedance bandwidth 5GHz) at 30GHz. Please suggest me easy method to design power divider for wideband applications.
Thank you very much
Kanhaiya Sharma
As you know, the null space of a matrix A is the set of vectors that satisfy the homogeneous equation Ax=0.
To find x (as the null space of A), I wrote two optimization models as below. I know, they are simple and straightforward and the solution may not be simply achievable but this is just my first basic idea.
--------------------------------------------------------------------------------------------------
1) Min Z=1
s.t.
sum(j , A(i,j) * x_null (j,m)) = zero(i,m);
where, Z is a dummy variable,
i*j is the dimension of A,
and m is assumed as a known column number of x.
But, the result is always x_null (j,m) = 0.
--------------------------------------------------------------------------------------------------
To deal with this problem, I modified (1) as below.
2) Max Z = sum((j,m) , x_null (j,m))
s.t.
sum(j , A(i,j) * x_null (j,m)) = zero(i,m);
Here, Z is the objective function.
In this model, the solver reports 'unbounded or infeasible'!
--------------------------------------------------------------------------------------------------
Note that, I let i<j make the number of equations less than the number of unknowns, and thus, the system is under-determined.
Any help would be highly appreciated!
I have been reading about performing sensitivity analysis of the solution of Linear Programming problem (calculating shadow prices, reduced costs and intervals within which the basic solution remains valid). It is clearly described on academical problems with 2 or 3 variables, but in fact, when tried to apply the same logic for real-life, scalable problem, I didn't get promising results. This is because only a few of variables values matters for me, while other are rather placed for another purposes (like changing hard constraints to soft ones etc). But all of them are taken into account when checking if basic solution has changed, hence the interval that is returned by a solver is a way more narrow than I want it to be.
Where can I find an example of real applied sensitivity analysis, if there is any?
Please give their appropriate cases.
Dear community,
I would like to request some references related to results on parameter-dependent Pareto Front, if there are any. I am interested in studying the behavior of the Pareto Front with respect to an external parameter, for multi-objective problems.
Thanks for any recommendation!
Best,
Nathalie
The attached picture is the result of a bi-objective optimization problem. Genetic algorithm was used. The "missed-out" band (i.e. approximately 41 to 43.5 in the vertical axis) is within the range of the respective objective function (in other words, there are values of design variables which result in values between 41 and 43.5 of the objective function).
The question is, is there any explanation for this discontinuity (physical or mathematical)? Or should I see this as a fault in my solution procedure?
It is worthy to add that I've carried out the procedure several times and with different optimization parameters (population, mutation fractions, etc), but the discontinuity seems to be always there...
hi everyone
i want to find a solution manual for this book
"Solution Manual For Applied Optimization with Matlab By P.Venkataraman"
can anyone help me with that please ?
i know that this book has a solution manual in Wiley Publications but it's an instructor manual and i don't have access to that so.....
this book has so many examples and i want to learn them but without a proper manual it's impossible to learn and code and i don't have enough time for that.
Thank you so much in advance
Hi All
I have the stress output of a structural analysis plotted against x ( x range is constant in all cases) that is a curve with minimums and maximums.
changing the model chracteristics ( stiffness ,etc) and doing a batch run, how could I code the optimization ?
preferably in Python
What are the links in their definitions? How do you interconnect them? What are their similarities or differences? ...
I would be grateful if you could reply by referring to valid scientific literature sources.
I'm trying to identify which approach would work best to select a set of elements that have different features that minimise a certain value. To be more specific, I might have a group of elements with Feature 1, 2, 3, 4 and another group with Feature 2, 3, 4, 5.
I'm trying to minimise the overall value of Feature 2 and 3, and I also need to pick a certain number of elements of each group (for instance 3 from the first group and 1 from the second).
From the research I did it seems that combinatorial optimization and integer programming are the best suited for the job. Is there any other option I should consider? How should I set up the problem in terms of cost function, constraints, etc.?
Many thanks,
Marco
I have 3 objectives in ILP model.The first has to be maximized and the second, and the third should be minimized.
I would like to compute the knee point of the generated Pareto front.
Didi you have an idea about the formula ?
thanks