Science topic

Iterative Methods - Science topic

Explore the latest questions and answers in Iterative Methods, and find Iterative Methods experts.
Questions related to Iterative Methods
  • asked a question related to Iterative Methods
Question
2 answers
Under the theme “Mathematics for All, Mathematics as a Path to Development,” one of the best examples of math's impact on daily life is its application in medical imaging, such as CT and CBCT scanners.
How do these devices work?
Mathematics is at the heart of it! When X-rays are emitted from a source, they pass through tissues and are captured by detectors. Along their path, these rays undergo absorption or attenuation, which depends on the properties of the tissue they traverse. Each ray's path represents a unique equation, where the unknowns are the attenuation values of the points it passes through.
As the source rotates around the object, it generates multiple paths, creating a system of n equations with n unknowns. Sophisticated software solves this complex system using advanced algorithms, such as the Radon Transform or iterative methods, to reconstruct the internal structure of the object as a detailed 3D image.
How does this contribute to development?
Mathematics is the backbone of such groundbreaking technologies, enabling precise diagnostics and effective treatment planning. This practical application of equations demonstrates how math can transform our world.
Now it’s your turn:
What other applications of mathematics in medicine and engineering do you think are underexplored? How can we further harness the power of math as a tool for development?
Let’s explore the incredible role of mathematics in everyday life together in this discussion!
Relevant answer
Answer
Follower@
  • asked a question related to Iterative Methods
Question
6 answers
The importance of finding a powerful numerical solution for the Laplacian matrix A in the Laplace and Poisson partial differential equations is obvious.
The classical conventional numerical procedure for solving the Laplace partial differential equation is based on the discretization of the 1D, 2D, 3D geometric space into n equidistant free nodes and on the use of the finite difference method FDM supplemented by the Dirichlet boundary conditions (vector b) to obtain a system of algebraic equations of order n prime in the matrix form A,
A . U(x,y,z,t) = b
The solution to the above equation is U(x,y,z,t) = A^-1.b which is often quite complicated since the matrix A is singular.
It is worth mentioning that common numerical iteration methods such as Gaussian elimination and Gauss-Seidel methods are complicated and require the use of ready-made algorithms such as those in Matlab or Python..etc .
We assume that there exists another SIMPLE statistical numerical solution expressed by:
A^-1=(I-A)^-1=A^0+A+A+A^2+....+A^N
Where A^0=I and N is the number of iterations or time steps dt.
As simple as that!
Relevant answer
Answer
The Laplacian theorem is an advanced and exhaustive form of the energy continuity equation. It can create new mathematics.
The new mathematics created by Laplacian's theorem is physico-mathematical rather than current pure mathematics and therefore more exhaustive and more understandable.
Recall the considerable success of Schrödinger's partial differential equation and quantum theory as a whole where it also relies on an advanced and exhaustive form of the energy continuity equation which is analogous to that of the proposed Laplacian mathematics.
The Laplacian partial differential equation (not to be confused with the famous Laplace PDE) describing the energy density U(x,y,z,t) has a beautiful and interesting form,
dU/dt)partial = α ∇^2 (U) + S
surprising is that the LPDE and Laplacian theorem are a kind of universal rules that can generate new mathematics.
The particularity of Laplacian's theorem is that it makes it possible to identify and prove or refute the majority of mathematical inconsistencies apparent at first glance.
Additionally, it can produce several never before known mathematical rules and correlations, in addition to solving the current time-dependent PDE in its most general form.
The classical mathematical process for numerically solving Laplacian PDEs is to advance via the following procedure in three consecutive steps:
i-Discretization into n free nodes,
ii- Treatment by finite difference method FDM then,
iii- Solve the resulting n first-order algebraic equations, which is quite difficult since the underlying matrix A is singular.
The above procedure results in a system of n first order algebraic equations in matrix form,
A U = b  . . . . . . . .
Where Anxn is the so-called Laplacian matrix and b is the vector of boundary conditions.
Which means that the solution of PDE 1 is,
U=A^-1 .  b   . . . . .
Obtaining the solution of n first order algebraic equations from last equation is quite difficult since the underlying matrix (A) is singular.
A revolutionary alternative statistical technique, based on B-matrix chains, which are a product of the Cairo techniques.
This technique is valid for both classical macroscopic physics such as the heat diffusion equation as well as modern microscopic quantum mechanics such as Schrödinger's PDE.
This technique completely neglects partial differential equations as if they never existed.
  • asked a question related to Iterative Methods
Question
3 answers
I have a large sparse matrix A which is column rank-defficient. Typical size of A is over 100000x100000. In my computation, I need the matrix W whose columns span the null space of A. But I do not know how to fastly compute all the columns of W.
If A is small-scale, I know there are several numerical methods based on matrix factorization, such as LU, QR, SVD. But for large-scale matrices, I can not find an efficient iterative method to do this.
Could you please give me some help?
Relevant answer
Answer
If the matrix is sparse then a sparse LU will enable you to compute the null space. (Row &/or column swaps (= pivoting) may be necessary or useful to keep memory and time costs low.) But otherwise this will become very expensive. Even just storing such a matrix would take (in double precision floating point) over 80GB. Any solver would be an out-of-core type of solver, but would be extremely expensive.
An alternative, if the matrix is real symmetric, is to use the Lanczos method or a variant such as the Lehoucq & Sorensen ARPACK method. More precisely, ARPACK is a restarted Arnoldi method. Then you look for when you have the zero eigenvalue (or simply a very small eigenvalue). The corresponding eigenvector(s) gives a basis for the null space. ARPACK is a semi-iterative method, and so there is a trade-off between the accuracy of the eigenvalues (and eigenvectors) and the number of iterations performed.
  • asked a question related to Iterative Methods
Question
3 answers
For large-scale matrix Aand vector v, I want to to compute R^{-1]v, where A=QR is the QR factorization of A. The explicit computation of QR and R^{-1} is unaffordable.
Is there any iterative method to approximate R^{-1]v implicitly?
Relevant answer
Answer
Hi,
Please take look at those attached files.
Best regards
  • asked a question related to Iterative Methods
Question
3 answers
I need some papers links in which we have a solution for system of first order odes using the Laplace variational iteration method (VIM)?
Relevant answer
Answer
I found this solution of the Lotka-Volterra system of first-order ODEs via VIM:
And here is a good overview of how to leverage a Laplace Transform when solving ODEs with VIM:
I hope you can put these together for LVIM.
  • asked a question related to Iterative Methods
Question
6 answers
Which controller is best in trajectory tracking, i am getting same response.
please share your experience
Relevant answer
Answer
Thanks sir, i will keep in touch
  • asked a question related to Iterative Methods
Question
9 answers
Suppose there is a linear system of equation Ax =b, x is required. A is a large symmetric matrix.
Which method is faster to compute x?
1. If I solve it through the calculation of inverse of A and compute x = A-1b.
2. If I solve the linear equations through some iterative method (like Gauss-elimination, Gauss-seidel).
Relevant answer
Answer
Hi Iman,
For large symmetric matrices, the Conjugate Gradient (CG) and Conjugate Gradient Least Squares (CGLS) methods are very effective, super fast, and require little memory to implement. The entire matrix A need not be loaded into memory, and the method can read columns and rows of matrix A to apply the algorithm. See this excellent attached book on inverse problems (Aster et al.) for more details.
Regards,
Hamzeh
  • asked a question related to Iterative Methods
Question
3 answers
I am working on a numerical problem that involves large banded matrices, I am trying to solve the problem using an ADI scheme however the problem is really sensitive to the acceleration parameters I use (i.e. most choices lead to $\rho(T_{w_n})>1$), the problem tends to converge fast for the first iterations but diverges afterwards.
I tried to use some of the values of $w_n$ I found in literature and the one provided in Ref. [1]. seems to be the closest fit for my problem but still leads to a $\rho(T_{w_n})>1$.
For small problems I can use the power iteration and the Simulated Annealing proposed by Koutserimpas et al. [2] (since their suboptimal optimization parameters don't seems to work for me and leads immediately to a diverging result), but it becomes impractical for larger ones.
In Ref. [2] a linear dependence on the discretization for $w_n$ is proposed but I am not sure if it is a general rule or a quirk in their simulations. Since I was working with preconditioner on a multigrid as an alternative to the ADI scheme I was curious to know if I could extrapolate a general relationship between the spectral radius of a matrix for a discretization of $\Omega_h$ and it's upscaled smoothed version for $\Omega_{2h}$.
Is the linear dependence found in Ref. [2] a coincidence for their problems? Or is there some theorem I can use?
[1] Hadley, G. Ronald. "Numerical simulation of reflecting structures by solution of the two-dimensional Helmholtz equation." Optics letters 19.2 (1994): 84-86.
[2] Koutserimpas, Theodoros T., Aristeides D. Papadopoulos, and Elias N. Glytsis. "Applicability and optimization of the alternating-direction-implicit iterative method for the 2-D finite-difference frequency-domain solution of scattering problems." IEEE Transactions on Antennas and Propagation 65.12 (2017): 7166-7173.
Relevant answer
Answer
This article might help you with you question
  • asked a question related to Iterative Methods
Question
4 answers
What are the most recommended publishers for submitting a book in the field of Iterative Methods for solving fractional Differential Equations?
Relevant answer
Answer
Thank you very much
  • asked a question related to Iterative Methods
Question
3 answers
BICGSTAB-FFT can be used in DDA because the special form of its interaction matrix, which makes the time complexity reduce from O(N^2) to O(NlogN). Iterative method such as bicgstab can also be used to solve the matrix in finite element method, but it seems FFT(fast fourier transform) can not be implemented in this case.
Does this mean, suppose with the same number of elements (or matrix size), finite element method will principally be more time consuming than DDA simply because the more time consuming matrix vector multiplication in bicgstab (suppose bicgstab runs the same number of iterations).
  • asked a question related to Iterative Methods
Question
4 answers
I have implemented a BiCGSTAB iterative solver, for a certain linear equation, the Jacobi preconditioner actually deteriorates the convergence rate. The matrix A is positive definite and nothing really special. Based on my understanding of BiCGSTAB, with a Jacobi preconditioner, it should be always easier to solve such a linear equation than a BiCGSTAB solver without a preconditioner. However, when I remove the preconditioner, it converges much faster.
So can anyone tell me what is going on here? Do you understand the BiCGSTAB correctly?
Relevant answer
Answer
First, BICGSTAB is a non-symmetric sparse matrix solver which works better than its competitors (GMRES, TFQMR, etc.) in general. Try to implement the restarted version BICGSTAB(p) where p is the order of last iterates the algorithm keeps to calculate the current iteration. Ideally, p=2 is fine as it converges in much fewer iterations with a marginal increase of cost/iteration.
Second, preconditioning is a key to accelerating any sparse iterative solver. However, this promise is only valid when you construct a suitable preconditioner method for your particular problem. There is no guarantee whatever that a preconditioner that nicely works for one class of physical problems will be suitable to others. You might test other kinds of preconditioning techniques such as ILU(0), ILUTP, AMG. But keep in mind that the best choice is often challenging.
  • asked a question related to Iterative Methods
Question
1 answer
Hi
Yes as it's, the system is not easy and requires so much work and tears. I will summarize my question to make it easier to understand. So I have system of coupled pdes. The unknown variables are dependent on time and space: du/dt=f(u'', u', u, v'', v', v, s'', s', s), dv/dt=f(u'', u', u, v'', v', v, s'', s', s), ds/dt=f(u'', u', u, v'', v', v, s'', s', s). So I have used finite difference method to discretize the derivative terms. However, since they are nonlinear, I should choose an iterative method to compute the solution. I was thinking to implement Picard method, but skeptical that the iteration will never reach the minimum tolerance between the solution and initial guesses. So which method would you recommend to use?
Thank you
Relevant answer
Answer
maybe working in cycles for u then v then s and for several space variables
in an ADI scheme. Then there remains ''only" one variable on the grid varying for
every ssubstep. and
for this one Newtons method? But more help may come from considering the physics behind in more detail. Are there some dominant variables / space components etc.?
  • asked a question related to Iterative Methods
Question
7 answers
I want to use LES to simulate the flow past a sphere at high Re by non-iterative PISO method, which is much faster than iterative method. However, the result of non-iterative PISO is far from the result in literatures, especially in Strouhal number. Can this method stable and accurate enough to apply to LES in turbulent flow? How can I improve the accuracy?
Relevant answer
Answer
Try the NITA based on fractional time step and compare to see if you have better solutions.
  • asked a question related to Iterative Methods
Question
4 answers
For example, Broyden’s method, Newton iteration method, ... .
I did not write much program with it. Also, I do not have much time.
I have attached some explanations.
Thank you
Relevant answer
Answer
The type of nonlinear system need be clarified. Is it system of scalar nonlinear equations? system of nonlinear BVPs of ODEs or PDEs? System of nonlinear IBVPs of ODEs or PDEs...?
In case of the usual system of nonlinear scalar equations, the matlab symbolic tool (command) ''solve'' may be used. Numerical methods are Newton method, Halley's method, Chebyshev's method....There are references with matlab codes using the iteration methods.
  • asked a question related to Iterative Methods
Question
4 answers
Nonlinear algebraic or transcendental scalar equations are usually solved by using iteration methods. Because some of them are so intricate to solve analytically or an approximate numerical solution is required. Nonlinear algebraic equations can occur during several modeling processes or when solving nonlinear BVPs using FDMs......
Relevant answer
Answer
I cover all of the best methods in this book, which will be free on Thursday (7/15) https://www.amazon.com/dp/B07FL7JR1J The software examples are free here http://dudleybenton.altervista.org/software/Nonlinear_Equations_examples.zip In answer to your question: It depends. Some methods are better than others. Some work well for one class of problems but not another class. In the literature you often find problems carefully selected to prove a point or make one method stand out above all the rest, which might make for an interesting paper but is not really helpful. I try to present a balanced approach.
  • asked a question related to Iterative Methods
Question
5 answers
I wonder that if anyone tried SOR method for unsteady problem discretized by finite difference or finite volume method. Does it converge faster AND solve faster when it is compared to Gauss-Seidel and Jacobi solver?
Relevant answer
Answer
SOR will converge faster than GS or Jacobi methods if the relaxation factor is chosen well. If, by diffusion equation, you mean Fourier's equation, a linear equation, then SOR will be perfectly fine. Easy to encode and not complicated. BiCGSTAB etc are powerful methods but are unnecessarily complicated for something like this. Quasi-Newton methods are for nonlinear equations. But if you have a nonlinear diffusion equation then these more powerful methods may well be the way to do it.
  • asked a question related to Iterative Methods
Question
3 answers
As is well known, iterative methods for solving linear systems such as Successive Over Relaxation and the like, are very attractive for solving many problems such as sparse matrices. These methods, in general, are formulated in the context of determined system in which the number of equations is equal to the unknowns. Now and for sake of simplicity, let us assume that we have one additional observation and We need to update the previous solution. In other words, now we have an over determined system with the provision of this additional observation. The question is how to include this observation to the to update the previously computed parameters. Indeed, the theory of parameter estimation provides a lot of guidelines to handle this task to get an optimal solution in the sense of least squares. But let us assume that we need to stick to the iterative approach for parameters. Then with assumption how we could handle the additional observation for parameters update and what kind of errors that we need to minimize.
Relevant answer
Answer
If you are looking for sparse solution to a linear system of arbitrary size m by n, you can apply the iterative method called " Orthogonal Matching Pursuit (OMP) algorithm ".
Also, Lasso Regression model can be used in such cases, which involves a regularization term in a L1 norm. For sparsity in solution, minimization in L1 or Linfinity norm is sought.
OMP algorithm generates a sparse solution based on minimization in L0 norm ( number of non zero entries in the vector)
  • asked a question related to Iterative Methods
Question
4 answers
What is the relationship between iterative methods and line search techniques?
Relevant answer
Answer
It is not always possible to guess more or less appropriate step size in steepest descent algorithms. In some cases it may be expedient to perform parabolic one-dimensional search.
The computational work for such search is usually neglectable comparing with obtaining of the gradient of minimizing function.
  • asked a question related to Iterative Methods
Question
3 answers
I'm new DIANA user and now I am working in numerical analysis of reinforcement concrete infilled frames with DIANA.
So I have a problem when it comes to running the model. After results these errors appears
SEVERITY : FATAL
ERROR CODE: /DIANA/AP/LB40/0053
ERRORMSG.F: Divergence occurred, iteration method failed.
Use smaller load steps, different iteration procedure or different
control procedure. Restart from a saved FILOS file or rerun the
complete job.
SEVERITY : ABORT
ERROR CODE: /DIANA/LB/ER50/0098
ERRORMSG.A: 1 fatal error messages occurred during previous analysis.
Can't continue.
DIANA-JOB ABORTED
I let you check my model whether you can bring your experience in.
Thanks in advance
Relevant answer
Answer
In addition to the answer above, I wouold say that the method in numerical modeling is to begin first with everything with elasrtic behavior and to compare the results with an analytical study. It will help you as well to be sure the problem is in the model itself and not in the definition of the behavior of the elements.
For the hysteresis behavior, search the Equivalent Strut Model. Several hysteresis behavior exist.
  • asked a question related to Iterative Methods
Question
8 answers
How can i solve system of differential equation using picard iterative method?
Relevant answer
Answer
I suggest you Picard-Green and Picard-Mann methods which and newly introduced methods.
  • asked a question related to Iterative Methods
Question
4 answers
I am looking for a new iterative method or analytic-approximate method to solve ODE\PDE not the already known methods such as:ADM, VIM, HAM, HPM, DJM, DTM,,,,,but new method.
Best regards,
Relevant answer
Answer
please see Picard-Green and Picard-Green methods.
  • asked a question related to Iterative Methods
Question
11 answers
such as homotopy analysis method
Relevant answer
Answer
Picard-Green , Mann-Green and Ishikawa-Green methods are new and best.
  • asked a question related to Iterative Methods
Question
1 answer
He's Variational Iteration method is applied in my paper for obtaining an approximate and analytical solution of non-linear differential equation containing a non-linear term related to enzymatic reaction.
Relevant answer
  • asked a question related to Iterative Methods
Question
1 answer
I have read the article "A-Optimal Projection for Image Representation" ().
The authors used an iterative method to solve the optimization problem, i.e. obtain the transformation matrix A. But the number of iterations hasn't been discussed in the paper, and I fail to find the code of the paper thus not know the performance in practice.
How to determine the number of iterations of such a form of iterative method?
Relevant answer
Answer
Dear Shijie , To the best of my knowledge, most of the optimization methods are iterative, and the number of iteration should be quantified by try and error. It drastically depends on the problem you want to solve.
  • asked a question related to Iterative Methods
Question
3 answers
The proportional fitting procedure (IPFP,) is an iterative algorithm for estimating expected cell values [M_ijk] of a contingency table such that the marginal conditions are met.
For iterative computation of the M_ij, using MLE for the model, other Algorithm exist , and most common are for 2D while am interested in that of 3D, Assuming model of independence is equivalent to the model that all odds ratios equal one. I need references of examples showing manual calculation, and how to do more advance ones on R.
Relevant answer
Answer
Use table function; is you share the exact breaker you faced, probably I can help you out.
  • asked a question related to Iterative Methods
Question
6 answers
I am programming a code for an iterative method of solving a PDE (partial differential equations) problem under Matlab using the finite element method. The solution I get changes depending on the mesh I choose. I know that this is illogical. I wonder if anyone of you knows the reason of this problem.
Relevant answer
Answer
This is not completely illogical. The discrete solutions computed by a convergent method should all tend to the same limit as h (mesh size) --> 0, but for finite non zero values of h and different meshes it is very common that even very different results are obtained, especially if one is comparing meshes that are not completely homogeneous and are refined in a different way in different portions of the domain. The first step to check the correctness of what you are doing would be to take a series of meshes built each as a refinement of the previous one, so that h decreases each time by approximately a factor 2, and estimate the convergence rate of your approximation. If it does not correspond to the theoretical value, then you definitely have either an implementation error or you are studying a problem in which the convergence theory does not apply (for example because the solution is not regular), so depending on the case you either try to correct the code or rather to use a different benchmark.
  • asked a question related to Iterative Methods
Question
5 answers
Hello researchers,
I have been working recently on numerical analysis questions and came across a problem to which i am stuck. The image is attached.
Can you please help?
Thank you.
Relevant answer
Answer
The iteration function C(mi) depends on the constants phi and sigma. The standard sufficient condition for local convergence of the fixed point method is that |C'(mi)| < 1 in a neighborhood of the fixed point, which ensures what is usually called linear convergence. So the first step is to find values of phi and sigma such that this condition is true. I assume (but this is an interpretation, since this terminology is less standard) that by quadratic convergence it is meant that C'=0 at the fixed point and by cubic convergence that C'' =0 at the fixed point.
So this faster convergence can be achieved if one can find phi, sigma values such that these further sufficient conditions are true.
  • asked a question related to Iterative Methods
Question
3 answers
Does the convergence rate of multigrid depend on the total number of smoothing steps or on the number of pre and post smoothing steps seperately? What happen in the case where we have no pre-smoothing steps and more post -smoothing and viceversa.
Relevant answer
Answer
Dear Spyros, if you look to the iteration matrix of the two-grid method (TGM), then, using the same stationary method, you discover that the TGM iteration matrices in case of (s,t), (u,v) smoothing steps are similar as long as s+t=u+v. Therefore the asymptotic convergence speed is identical, but not the practical behavior, since the fastest convergence even at the initial iterations is observed when the global iteration matrix is the most close to a normal matrix..... In many sitiations, coming back at your example, (1,1) is better than (0,2) or (2,0).... But to prove it is somehow tricky.
  • asked a question related to Iterative Methods
Question
1 answer
The difficulty is that the observation data is not simultaneous, thus a direct empirical estimation is not feasible. Desroziers has proposed an iterative method based on posterior residual diagnostic but its convergence can be difficult to obtain, especially in a highly non linear dynamical system (e.g. Lorenz)
I am wondering if other methods have been tested, like a regression (observation data -> matrix) or an optimization modelling (find the matrix that minimize the output error in a dynamical system)
  • asked a question related to Iterative Methods
Question
3 answers
Dear the experts,
I have seen a lot of studies using deep learning into tomographic imaging, and improving the computational time of it.
I wonder if there is another measure for comparing computational costs.
If I am right, I found that FLOPS is widely used metric in deep learning models.
But in the iterative methods, it would not be appropriate because there can be the computations beyond the simple multiplication and summation.
If there is a good way to, I would like to ask you about it.
Thank you in advance
Relevant answer
Answer
  • asked a question related to Iterative Methods
Question
4 answers
BFGS algorithm is known to be an iterative method for solving unconstrained nonlinear optimization problems. Will different presentations of objective functions affect the applicability/feasibility/convergence of the BFGS algorithm?
On the other hand, SQP is known to be an iterative method for solving constrained nonlinear optimization problems. Will different presentations of objective functions and/or constraints affect the applicability/feasibility/convergence of the SQP?
What are the rationales behind?
Relevant answer
Answer
One of many glitches is that if the problem is ill-posed, then the subproblem can be infeasible.
  • asked a question related to Iterative Methods
Question
1 answer
I am writing a code for solving Eliashberg equations(self-consistant) by simply iterating method( x=f(x)), but not getting convergence.
Can someone suggest some mixing or any oyher method for getting fast convergence?
Relevant answer
Answer
Try bisection method. You can combine it with iterating method xn+1=f(xn) if xn+1 falls into the interval where there is a nul-point. If it doesn't, then bisect.
  • asked a question related to Iterative Methods
Question
2 answers
I'm working on iterative phase tomography which combines the phase retrieval and image tomography together, while few papers could be found till now. Can you recommend me some of them?
Thanks!
Relevant answer
Answer
Yizhou Lu , Thanks for your recommendation.
I have skimmed over the book , while the book is introducing the iteration between x-ray and matters, it's more related to the application of x-ray in material science. The last chapter is the absorption and phase contrast imaging of x-ray, but there seems no phase retrieval related content in the book. Anyway, thanks you!
  • asked a question related to Iterative Methods
Question
3 answers
How can we apply adomian decompostion method when there is no linear term in differential eqation
Relevant answer
Answer
In this paper, linear and nonlinear boundary value problems for fourth-order fractional integro- differential equations are solved by Variational iteration method (VIM) and Adomian decomposition method (ADM). The fractional derivative is considered in the Caputo sense . The solutions of both problems are derived by infinite convergent series . Numerical example are presented to illustrate the efficiency and reliability of two methods. General Terms Numerical solutions, Fractional integro-differential equations. Keywordsractional integro-differential equations, Caputo fractional
  • asked a question related to Iterative Methods
Question
1 answer
These two methods (1) Picard Iterative Method(PIM)
(2) Temini and Ansari Method (TAM)
seen almost similar both in principles and applications especially when they are applied to initial-value problems. Can someone tell me the differences ?
Relevant answer
Answer
VIM, HAM, ADM and so on are all equivalent methods to PIM. The general use of Lagrange multipliers, as in VIM, underlies all these methods. For instance, ADM and PIM can both be regarded as variants of VIM, see "A Unification of the Concepts of the Variational Iteration, Adomian Decomposition and Picard Iteration Methods; and a Local Variational Iteration Method" by Xuechuan Wang, Satya N. Atluri . Each of these methods have merits and defects concerning order and speed of convergence.
TAM should provide excellent approximations to the solution of non-linear equation with low computational time because w.r.t. a method based on Lagrange multipliers it shouldn't the compute the value that solves the boundary condition. TAM is apparently not based on Lagrange multipliers and if so it uses low computational time w.r.t. PIM. But if one proves that they are related, nothing done.
I hope to have been helpful (at least a little),
Fabrizio
  • asked a question related to Iterative Methods
Question
4 answers
Dear friends
I have a serious problem in order to regenerate the results of the attached paper. I do exactly mentioned method (FEM) in the paper but I don't get correct results.
This paper about gas face seal and compressible zeroth and first order perturbed Reynolds equation. I Solved zeroth order with FEM and FDM correctly and I got same results that bring out in the paper (Opening force for various pressure ratio) and I solve first order equation with 1 pressure ratio (figure 5 and 6) (I am thankful to you if look at the paper), but in other figure I can't get the paper's results. I try many method to solve linear system of equation (Ax=b) such as direct method, iterative method (i.e. gauss-sidel, cgs, gmres, bicg, pcg ...) but I failed, also I try many network grid and there is no achievement.
So what should I do?
I really know my question is general, but I really don't know about the errors.
Relevant answer
Answer
Thanks dear Ryan Vogt and Debopam Ghosh
  • asked a question related to Iterative Methods
Question
1 answer
As newton raphson method find one root it a time while simultaneous iterative methods approximate all root at time are the above procedure hold for it?
Relevant answer
Our procedure looks at a numeric method as the iteration of an application N of a space X in itself. The space X is divided into several basins of attraction some of which correspond to the roots of the equation. The use of dynamics can help to find an adequate set of initial points so that the iteration process ends up reaching all the roots of the equation, but it is not a technique that exclusively has this objective. When the map N is a branched covering, it has special properties, for instance, the associated Julia set J(N) has many properties of symmetry.
  • asked a question related to Iterative Methods
Question
4 answers
Suppose we've found the dominant e.value & corresponding e.vector by power method and Rayleigh quotient. Then what about the others?!
Relevant answer
Answer
We outline five such iterative methods to find all the eigenvalues & eigenvectors
the power iteration method, the shifted inverse iteration
method, the Rayleigh quotient method, the simultaneous iteration method, and the QR method.
  • asked a question related to Iterative Methods
Question
3 answers
I am currently working on a project in which we aim to produce a thermo-labile receptor. in the course of my preliminary research I came across the B FITTER program which calculates the amino acid B-factor as an average of the B-factors of all of its atoms excluding hydrogen, then generates an output file with a ranked classification of all the residues according to the B-factor. (described here: Reetz MT, Carballeira JD. 2007. Iterative saturation mutagenesis (ISM) for rapid directed evolution of functional enzymes. Nat Protoc 2:891–903.)
My problem is that some of the PDB files generate a proper output file, while others turn up a blank file (example: 3K90 produces a proper file, while 3KB0 doesn't). Could this be caused by a difference in the PDB file format? is there a way around this issue?
Many thanks in advance,
Dan.
Relevant answer
Answer
Although I am not familiar with B FITTER, my guess would be the following:
3K90 contains only isotropic B-factors, while 3KB0 contain the more informative anisotropic B-factors. Isotropic B-factors are listed in the ATOM records of a PDB file, while anisotropic B-factors have their own ANISOU records. Look at the pdb files in a text editor to see the difference, and check whether other problematic files also use anisotropic B-factors. see https://www.phenix-online.org/documentation/dictionary.html for a definition/image of the difference between isotropic and anisotropic B-factors.
  • asked a question related to Iterative Methods
Question
4 answers
Just for test; I applied a fixed (known) transformation to a point cloud, afterwards I applied ICP to estimate my initial transform. I noticed that the values obtained from the estimation with ICP were not same with my initial transform. If ICP doesn't provide a unique solution, how can we ensure its accuracy?
**Edit-1:
I compared both initial and estimated transforms in terms of Euler's angles rather than conventional Rotation matrices.
Relevant answer
Answer
Most ICP algorithms do NOT give global solution but just the best local minimum so that's why it's important to give them an initial transformation that will be close to the real transformation.
Software like Meshlab will usually let the user specify 3-4 equivalent points on both mesh/point clouds and start aligning these corresponding points first before using ICP...
  • asked a question related to Iterative Methods
Question
5 answers
There are two computational aspects in computing the zeros of orthogonal polynomials(Gauss quadrature nodes)
1: Eigenvalue approach
2: Root finding Iterative methods like as Newton's method.
Which one of the above schemes do not suffer from round-off errors?
Which one of these two technique is suitable specially for large values of
approximation degree?
Best.
Relevant answer
Answer
I think that both methods have self-correction, which means that the round-off errors do not accumulate. These are iterative methods, and on the next iteration, the round-off errors of the current iteration will be corrected.
Some (obsolete) methods for eigenvalues include finding the polynomial roots. The modern advanced methods for eigenvalues are based on the matrix decomposition. Furthermore, if the polynomial roots are needed, the square matrix is built from the polynomial coefficients, such that is eigenvalues are polynomial roots (and Newton method is avoided).
As I mentioned, the round-off errors are not crucial because both methods are self-correcting, but the eigenvalue method has some other advantages. In particular, unlike the Newton method, the initial guess is not needed.
  • asked a question related to Iterative Methods
Question
4 answers
Hi,
I am trying to implement MLEM algorithm for SPECT image reconstruction in 3D. But I am confused with probability matrix. How can I calculate probability matrix?
Thanks in advance
Manu
Relevant answer
Answer
Analytical calculation+ PSF, Ref: Accurate image reconstruction based on standard-Gaussian-model fitted system matrix in multipinhole small animal SPECT imaging
  • asked a question related to Iterative Methods
Question
5 answers
I need to do iterative more than 1000 simulation of a job with just small change in each run, so i do not want to check for warnings and errors in each run, as it is taking A LOT OF TIME, and directly just submit and run the analysis each time..
Relevant answer
Answer
First run a datacheck only, then run the job from the command line using the option
continue.
See the execution section (3.2.2) in the manual
  • asked a question related to Iterative Methods
Question
3 answers
Large scale sparse matrix iterative solution- 1 million Order level
Relevant answer
Answer
see this "A large-scale sparse matrix multiplication method
based on streaming matrix to GPUs" it useful
  • asked a question related to Iterative Methods
Question
6 answers
For sure, this simple mapping has been explored:
Y_{n+1} = y (1+Y_n)/(1+ Y_n +y).
For y < -0.25 it oscillates. Can anybody point me to a study? 
Relevant answer
Answer
The iteration scheme
x_{n+1} = 1/2(-1/x_n + x_n)
which is the solution of the equation x^2 + 1 = 0
converges for all x_0 except real.
I find that curious.
  • asked a question related to Iterative Methods
Question
6 answers
Hi everyone!
Is someone have sucessfully implemented a DG code ( in matlab, fortran, python or alike... not in C/C++) for simulate highly convective viscous incompressible flows?
I did not have success in doing it.
Let me explain:
I have used the DG code by Hesthaven & Warburton (J.S. Hesthaven, T. Warburton, Nodal Discontinuous Galerkin Methods - Algorithms, Analysis, and Applications, Spinger, (2008)). This code only works under low Reynolds number flows ( at maximum about 100 Re ). So I begun to try extend this code so it runs high-Reynolds number flows.
I adopted the ideas from Krank eta al article (A high-order semi-explicit DG solver for 3D incompressible flow with application to DNS and LES of turbulent channel flow - arXiv ). These authors made a C++ DG code ( using projection method "a la Karniadakis" ) and it seems they was successfull in simulating high-Reynolds number viscous incompressible flows. Until now, I have implemented the new convective-step part ( from the three parts that compounds the projective method ) and kepting all othes program's subroutines untouched. Unfortunately I dont had sucess. The program blows up, even running low Reynolds number flows!? It seems strange to me, for I would expect this unique program's change do not affected the entire program behaviour, as I have changed the old convective step ( that works for just low Rynolds number flows) by a new (better!) one, that was conceived to run highly convective flows!?
What is happening?
Can someone help-me?
Relevant answer
Answer
Thank you Karakus.
In fact, just one partial integration is applied in the articles of Krnak et al and Fehn et al ( for example...) "On the stability of projection methods for the incompressible N-S equations based on high-order DG discretizations" , also more rich cubature method is advised. The issue concerning small 'dt' ( reported by Ferrer, Krank  and others ... ) has been taken into account. In fact the base program of Warburton & Hesthaven evaluate satisfactory 'dt' for low-Reynolds flows.
My substitute m-function for original convective step do uses only one partial integration and element-by-element calculation of Lax-Friedrichs' stabilization parameter. This strategy was conceived targeting high-Re incompressible viscous flows. But my program blows up even at low-Re !
Now I will , also, atempt to apply a rich cubature method to volume integral .
Let's see what happens.
  • asked a question related to Iterative Methods
Question
7 answers
hello everyone;
i'm trying to implement lattice FIR filter and RII lattice filter in c.
my question is how to make the first g(n-1), some one can help me please! i know that z-1 represente a delay time so how can i program it ?, I thought to take each time the previous sampler, but it will not work for the first samples because we will have negative values
you can find the algorithm here
Relevant answer
Answer
thank you for all your answers, i'm getting the solution :)
  • asked a question related to Iterative Methods
Question
1 answer
I'm trying to solve the multi-objective optimization problem 
%% optimisation multiobjective en utilisant la méthode de pondération%%%%%
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
function [f] = objfun2(x,w1,w2)
% X1 : Entrée en électricité
% X2 : Entrée en gaz naturel
% X3 : Entrée en biomasse
% X4 : biogaz à exportation
% X5 : Facteur split 1
% X6 : Facteur split 2
% X7 : Facteur split 3
% X8 : Puissance échangée avec l'élement du stockage d'hydrogène
% X9 : Energie stockée
% D_ch :Demande en chaleur
% D_ele : demande en électricité
% Pour le scenarion 1:
n_ee = 0.90; % Trasformateur
n_EL_H2 = 0.85 ; % Eléctrolyseur
n_EL_TH =0.15 ; % Eléctrolyseur
n_GB_TH = 0.90; % Gas boiler
n_FC_EL = 0.55; % Fuel cell
n_FC_TH = 0.45; % Fuel cell
n_RF_H2 =0.7 ; % Reformer
n_FR_GN = 0.7; % fermenteur
n_H2_CH =0.8 ; % charge H2
n_H2_DE =0.8; % Décharge H2
C1 =0.8; % Cout spécifique Transformateur
C2 =0.20; % Cout spécifique chauddière à gaz
C3 =0.6; % Cout spécifique reformer
C4 =2; % Cout spécifique fermenteur
C5 =2; % Cout spécifique électrolyseur
C6 =2; % Cout spécifique pile à combustible
C7 =2; % Cout spécifique élément du stockage
f1=0;
for j=1:52
f1 = f1 + x(1)*C1+x(2)*x(5)*C2+(x(2)*(1-x(5))+(x(3)*n_FR_GN)*(1-x(6)))*C3+x(3)*C4+(x(1)*n_ee*(1-x(7)))*C5+((x(1)*n_ee*(1-x(7)))+(x(1)*(1-x(5))+x(3)*n_FR_GN*(1-x(6)))*n_FR_GN)*C6+x(8)*C7;
end
e1 =0.8; % facteur d'émission lié à l'usage d'électricité du réseau
e2 =0.20; % facteur d'émission lié à l'usage du gaz naturel dans la chadière
e3 =0.6; % facteur d'émission lié à l'usage du gaz naturel pour le réformage
e4 =2; % facteur d'émission lié à l'usage du la biomasse dans le fermentreur
f2=0;
for j=1:52
f2 = f2 + e1* x(1+9*(j-1))+ e2* x(2+9*(j-1))*x(5) + e3* x(2+9*(j-1))*(1-x(5+9*(j-1)))+e4*x(3+9*(j-1));
end
f = w1*f1+w2*f2;
end
%% Mes contraintes d'égalité et d'inégalité %%
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
% Mes contraintes d'égualité:
% Définir mes variables :
% X1 : Entrée en électricité
% X2 : Entrée en gaz naturel
% X3 : Entrée en biomasse
% X4 : biogaz à exportation
% X5 : Facteur split 1
% X6 : Facteur split 2
% X7 : Facteur split 3
% X8 : Puissance échangée avec l'élement du stockage d'hydrogène
% X9 : Energie stockée
% D_ch :Demande en chaleur
% D_ele : demande en électricité
function [c,c_eq] = mycon3(x,vout)
% Pour le scenarion 1:
n_ee = 0.90; % Trasformateur
n_EL_H2 = 0.85 ; % Eléctrolyseur
n_EL_TH =0.15 ; % Eléctrolyseur
n_GB_TH = 0.90; % Gas boiler
n_FC_EL = 0.55; % Fuel cell
n_FC_TH = 0.45; % Fuel cell
n_RF_H2 =0.7 ; % Reformer
n_FR_GN = 0.7; % fermenteur
n_H2_CH =0.8 ; % charge H2
n_H2_DE =0.8; % Décharge H2
% c_eq1 = x(1)*(n_ee*x(7)+n_ee*(1-x(7))*n_EL_H2) + x(2)*(1-x(5))*n_FR_GN*n_FC_EL + x(3)*(n_FR_GN*(1-x(6))*n_FC_EL*n_RF_H2);
% c_eq2 = x(1)*(n_ee*(1-x(7))*n_EL_TH+n_ee*(1-x(7))*n_EL_H2) + x(2)*(x(5)*n_GB_TH+n_RF_H2*(1-x(5)))+ x(3)*((1-x(6)*n_RF_H2*n_FR_GN) +x(8);
% c_eq3 = x(3)* n_FR_GN * x(6);
for j = 1:52
c_eq(j) = vout(j)-x(1+9*(j-1))*(n_ee*x(7+9*(j-1))+n_ee*(1-x(7+9*(j-1)))*n_EL_H2) - x(2+9*(j-1))*(1-x(5+9*(j-1)))*n_FR_GN*n_FC_EL - x(3+9*(j-1))*(n_FR_GN*(1-x(6+9*(j-1)))*n_FC_EL*n_RF_H2);
end
for i = 1:52
if x(8+9*(i-1)) >= 0
eh = n_H2_CH;
else
eh = 1/n_H2_DE;
c_eq(52+i) = vout(52+i)-x(1+9*(i-1))*(n_ee*(1-x(7+9*(i-1)))*n_EL_TH+n_ee*(1-x(7+9*(i-1)))*n_EL_H2) - x(2+9*(i-1))*(x(5+9*(i-1))*n_GB_TH+n_RF_H2*(1-x(5+9*(i-1))))- x(3+9*(i-1))*((1-x(6+9*(i-1))*n_RF_H2*n_FR_GN) +x(8+9*(i-1)))*eh;
end
for j = 1:52
c_eq(104+j) = x(4+9*(j-1))-x(3+9*(j-1))* n_FR_GN * x(6+9*(j-1));
end
% contraintes sur l'évolution de la charge des stockages
if x(8) >= 0
eh = n_H2_CH;
else
eh = 1/n_H2_DE;
end
c_eq(157) = x(9) - x(468) + 0.05 - eh*x(8);
for j=2:52
if x(8+9*(j-1)) >= 0
eh = n_H2_CH;
else
eh = 1/n_H2_DE;
end
c_eq(156+j) = x(9+9*(j-1)) - x(9+9*(j-2)) + 0.05 - eh*x(8+8*(j-2));
end
c = [];
end
%% Optimimisation %%%% Optimimisation %%
clear allclc
%% scénario demande en électricité et en chaleur par les batiments A-D
        %demande en Electricité batiments A B C D        scenario= [1   208816.92   164334.53   63948.44   14547.17;    2   208816.92   164334.53   63948.44   14547.17;  3   208816.92   164334.53   63948.44   14547.17;    4   208816.92   164334.53   63948.44   14547.17;    5   208816.92   164334.53   63948.44   14547.17;    6   208816.92   164334.53   63948.44   14547.17;    7   208816.92   164334.53   63948.44   14547.17;    8   208816.92   164334.53   63948.44   14547.17;    9   208816.92   164334.53   63948.44   14547.17;   10   208816.92   164334.53   63948.44   14547.17;    11   208816.92   164334.53   63948.44   14547.17;    12   208816.92   164334.53   63948.44   14547.17;    13   208816.92   164334.53   63948.44   14547.17;    14   208816.92   164334.53   63948.44   14547.17;    15   208816.92   164334.53   63948.44   14547.17;    16   208816.92   164334.53   63948.44   14547.17;    17   208816.92   164334.53   63948.44   14547.17;    18   208816.92   164334.53   63948.44   14547.17;    19   208816.92   164334.53   63948.44   14547.17;    20   208816.92   164334.53   63948.44   14547.17;    21   208816.92   164334.53   63948.44   14547.17;    22   208816.92   164334.53   63948.44   14547.17;    23   208816.92   164334.53   63948.44   14547.17;    24   208816.92   164334.53   63948.44   14547.17;    25   208816.92   164334.53   63948.44   14547.17;    26   208816.92   164334.53   63948.44   14547.17;    27   208816.92   164334.53   63948.44   14547.17;    28   208816.92   164334.53   63948.44   14547.17;    29   208816.92   164334.53   63948.44   14547.17;    30   208816.92   164334.53   63948.44   14547.17;    31   208816.92   164334.53   63948.44   14547.17;    32   208816.92   164334.53   63948.44   14547.17;    33   208816.92   164334.53   63948.44   14547.17;    34   208816.92   164334.53   63948.44   14547.17;    35   208816.92   164334.53   63948.44   14547.17;    36   208816.92   164334.53   63948.44   14547.17;    37   208816.92   164334.53   63948.44   14547.17;    38   208816.92   164334.53   63948.44   14547.17;    39   208816.92   164334.53   63948.44   14547.17;    40   208816.92   164334.53   63948.44   14547.17;    41   208816.92   164334.53   63948.44   14547.17;    42   208816.92   164334.53   63948.44   14547.17;   43   208816.92   164334.53   63948.44   14547.17;    44   208816.92   164334.53   63948.44   14547.17;    45   208816.92   164334.53   63948.44   14547.17;    46   208816.92   164334.53   63948.44   14547.17;    47   208816.92   164334.53   63948.44   14547.17;    48   208816.92   164334.53   63948.44   14547.17;    49   208816.92   164334.53   63948.44   14547.17;    50   208816.92   164334.53   63948.44   14547.17;    51   208816.92   164334.53   63948.44   14547.17;    52   208816.92   164334.53   63948.44   14547.17];  
        %demande en chaleur batiments A B C D          scenario2 = [1 1427594.60   1284109.59   523483.67   190031.15   2 982931.18   872385.05   368411.96   134151.09   3 1135105.91   966735.18   413340.25   149325.03   4 1495313.57   1299812.17   503369.62   188862.93   5 1325288.52   1059053.42   417684.96   151609.55   6 995935.34   729842.68   326075.69   110981.31   7 895859.46   748401.59   324866.08   114330.22   8 1064518.63   897955.07   376706.41   134196.02   9 884647.89   701571.18   300466.56   108722.74   10 800896.54   531547.79   265629.86   93068.54   11 690779.38   574932.49   284292.38   95440.39   12 630955.39   491976.27   218766.20   71651.09   13 451613.74   302690.85   182270.61   59536.17   14 282668.92   188690.11   127527.22   39082.41   15 365144.19   231710.71   164852.26   49844.24   16 375957.22   212325.89   141524.11   44132.92   17 300533.75   182593.74   146811.82   44003.12   18 279095.43   189308.38   150077.76   42990.65   19 139571.14   69729.07   83462.93   23753.89   20 199881.76   83818.97   115863.14   31542.06   21 85175.05   52589.89   70917.57   18691.59   22 93955.75   44086.60   45412.13   12980.27   23 131220.95   0             15552.10   3115.26   24 108755.75   0             17107.31   2076.84   25 116913.61   59.38       15552.10   2076.84   26 53402.44   0           12441.68   3115.26   27 47767.21   0           12441.68   2336.45   28 24543.93   0           12441.68   0     29 110250.54   0           12441.68   1038.42   30 151976.55   0           12441.68   3115.26   31 69422.12   0             12441.68   3115.26   32 86918.05   0             12441.68   3115.26   33 34192.03   0           12441.68   4672.90   34 20738.80   0             12441.68   6230.53   35 39519.77   0             12441.68   4153.69   36 68539.80   0             12441.68   3115.26   37 55234.83   0             12441.68   3115.26   38 292989.19   0             30482.12   3115.26   39 187184.72   126185.65   18662.52   23260.64   40 183608.26   143900.02   154789.13   48286.60   41 287919.61   240537.02   122083.98   56464.17   42 677292.12   605297.97   280715.40   112409.14   43 884163.86   742486.71   368584.76   123831.78   44 661414.79   501422.13   273999.72   89304.26   45 775772.10   632838.22   276827.37   110072.69   46 714229.28   673670.97   298859.51   108870.31   47 1109646.28   1082949.28   386823.13   169781.93   48 1387665.61   1405851.38   538102.64   214330.22   49 1178721.87   1101958.19   481832.32   169781.93   50 1500936.04   1296977.48   495889.80   213914.85   51 1437692.34   1143034.01   554765.61   182717.05   52 1504741.00   1394642.85   497874.55   192951.71];
for i = 1:52    vout(i) = scenario(i,2)+scenario(i,3)+scenario(i,4)+scenario(i,5);end 
for i = 1:52    vout(52+i) = scenario2(i,2)+scenario2(i,3)+scenario2(i,4)+scenario2(i,5);end
% contraintes d'inégalités linéaires%  A*X  <= bA = [];b = [];
% contraintes d'égalités linéaires%  Aeq*X  = BeqAeq = [];beq = [];
% initialisation du vecteur des variables%x0 = [1;2;0.5;0.5]; 
%  mes limites inférieure et supérieures : 
mat1 = zeros(1,468);
for i=1:52    mat1(1,8+9*(i-1)) = -10000;end
lb = mat1;
mat2 = zeros(1,468);for i = 1:52      mat2(1,1+9*(i-1))= 20000;    mat2(1,2+9*(i-1))= 20000;    mat2(1,3+9*(i-1))= 20000;    mat2(1,4+9*(i-1))= 20000;    mat2(1,5+9*(i-1))= 1;    mat2(1,6+9*(i-1))= 1;    mat2(1,7+9*(i-1))= 1;    mat2(1,8+9*(i-1))= 10000;    mat2(1,9+9*(i-1))= 30000;end   ub = mat2;
for j=1:52    x0(1+9*(j-1))=100;       x0(2+9*(j-1))=100;      x0(3+9*(j-1))=100;      x0(4+9*(j-1))=100;       x0(5+9*(j-1))=100;    x0(6+9*(j-1))=100;     x0(7+9*(j-1))=100;      x0(8+9*(j-1))=100;        x0(9+9*(j-1))=100;end
for i = 0:0.02:1        w1 = i;    w2 = 1-i;    % -----X = FMINCON(FUN,    X0,A,B,Aeq,Beq,LB,UB,NONLCON,OPTIONS)
options=optimset('Algorithm','interior-point','MaxFunEvals',10000000);
[x,fval] = fmincon(@(x) multiobj(x),x0,A,b,Aeq,beq,lb,ub,@(x) mycon3(x,vout),options,w1,w2);
end
someone could help me with the problem, Matlab tells me Failure in initial user-supplied objective function evaluation. FMINCON cannot continue ?
Relevant answer
Answer
can u give code in english in mtalb file and please send documnetation for me to check
  • asked a question related to Iterative Methods
Question
3 answers
I have currently used the Krylov and Multifrontal methods in order to solve disperse linear systems. Next figure shows the characteristic curves number of nodes vs CPU time of the Krylov and Multifrontal methods. Does anyone know about direct or iterative methods for decrementing the computation time in the solution of large sparse linear systems? Which one?
Relevant answer
  • asked a question related to Iterative Methods
Question
5 answers
I am trying to solve a large system of linear equations AX=B, where A is a Jacobean matrix, non-symmetric, very large (about 18000*18000), sparse matrix. Which method is the fastest method for it?
Jacobean matrix is ordered in a way that each of the four parts of the Jacobean matrix is block diagonal.
Relevant answer
Answer
Dear Farnaz
I am too, working on structural analysis and resolving matrix and algebraic diff eqns, and if do you have any questions about, probably I can help
Sincerly
  • asked a question related to Iterative Methods
Question
9 answers
I want to minimize the error function, which I defined as the sum of the square distances between data points and a given numerical solution. My error function then solely depends on the ODE parameters. Which iterative optimization method should I use in order to minimize the error function in the parametric space? My ODEs are nonlinear. I would like a solution based on an open source software such as Octave.
Relevant answer
Answer
See an example of ODE fitting at www.civilized.com/pdffiles/tumor.pdf
  • asked a question related to Iterative Methods
Question
10 answers
Dear all,
I am running simulations in OpenFOAM with icoFoam. But after a few time steps, the Courant number increases to a very large value. I wonder if anyone know the possible reasons.
Time = 0.0125
Courant Number mean: 0 max: 0
DILUPBiCG: Solving for Ux, Initial residual = 1, Final residual = 3.14067e-11, No Iterations 18
DILUPBiCG: Solving for Uy, Initial residual = 0, Final residual = 0, No Iterations 0
DICPCG: Solving for p, Initial residual = 1, Final residual = 9.96628e-08, No Iterations 246
time step continuity errors : sum local = 2.62573e-13, global = 2.24611e-24, cumulative = 2.24611e-24
DICPCG: Solving for p, Initial residual = 0.00826593, Final residual = 9.75861e-08, No Iterations 216
time step continuity errors : sum local = 5.1487e-13, global = 4.59247e-24, cumulative = 6.83858e-24
ExecutionTime = 0.26 s ClockTime = 0 s
Time = 0.025
Courant Number mean: 5.34236e-06 max: 0.00488047
DILUPBiCG: Solving for Ux, Initial residual = 0.754291, Final residual = 8.55034e-11, No Iterations 16
DILUPBiCG: Solving for Uy, Initial residual = 0.330802, Final residual = 6.98711e-11, No Iterations 16
DICPCG: Solving for p, Initial residual = 0.420872, Final residual = 9.52255e-08, No Iterations 265
time step continuity errors : sum local = 5.26749e-13, global = -1.48591e-23, cumulative = -8.02052e-24
DICPCG: Solving for p, Initial residual = 0.0611023, Final residual = 9.76153e-08, No Iterations 261
time step continuity errors : sum local = 7.56513e-13, global = -8.53319e-24, cumulative = -1.65537e-23
ExecutionTime = 0.29 s ClockTime = 0 s
Time = 0.0375
Courant Number mean: 1.02253e-05 max: 0.0113322
DILUPBiCG: Solving for Ux, Initial residual = 0.528265, Final residual = 5.87478e-11, No Iterations 17
DILUPBiCG: Solving for Uy, Initial residual = 0.266124, Final residual = 9.26735e-11, No Iterations 17
DICPCG: Solving for p, Initial residual = 0.538462, Final residual = 8.96817e-08, No Iterations 287
time step continuity errors : sum local = 1.26773e-12, global = 6.13145e-22, cumulative = 5.96591e-22
DICPCG: Solving for p, Initial residual = 0.579184, Final residual = 9.6864e-08, No Iterations 280
time step continuity errors : sum local = 3.73869e-12, global = -6.34438e-22, cumulative = -3.78469e-23
ExecutionTime = 0.33 s ClockTime = 0 s
Time = 0.05
Courant Number mean: 4.4832e-05 max: 0.370209
DILUPBiCG: Solving for Ux, Initial residual = 0.390239, Final residual = 6.12355e-11, No Iterations 18
DILUPBiCG: Solving for Uy, Initial residual = 0.312606, Final residual = 1.82604e-11, No Iterations 18
DICPCG: Solving for p, Initial residual = 0.916305, Final residual = 9.79778e-08, No Iterations 292
time step continuity errors : sum local = 3.3445e-11, global = -1.1319e-21, cumulative = -1.16975e-21
DICPCG: Solving for p, Initial residual = 0.7854, Final residual = 9.34491e-08, No Iterations 282
time step continuity errors : sum local = 1.78264e-10, global = -2.58022e-20, cumulative = -2.6972e-20
ExecutionTime = 0.36 s ClockTime = 0 s
Time = 0.0625
Courant Number mean: 0.00188543 max: 15.577
DILUPBiCG: Solving for Ux, Initial residual = 0.344186, Final residual = 2.53233e-11, No Iterations 19
DILUPBiCG: Solving for Uy, Initial residual = 0.325652, Final residual = 1.0124e-11, No Iterations 19
DICPCG: Solving for p, Initial residual = 0.887083, Final residual = 9.42448e-08, No Iterations 295
time step continuity errors : sum local = 1.40168e-09, global = -4.72842e-18, cumulative = -4.75539e-18
DICPCG: Solving for p, Initial residual = 0.968391, Final residual = 9.63933e-08, No Iterations 290
time step continuity errors : sum local = 1.7746e-08, global = 1.16238e-17, cumulative = 6.86843e-18
ExecutionTime = 0.39 s ClockTime = 0 s
Time = 0.075
Courant Number mean: 0.535059 max: 1724.54
DILUPBiCG: Solving for Ux, Initial residual = 0.642258, Final residual = 0.0046226, No Iterations 1001
DILUPBiCG: Solving for Uy, Initial residual = 0.966405, Final residual = 0.00529744, No Iterations 1001
DICPCG: Solving for p, Initial residual = 0.892676, Final residual = 0.036825, No Iterations 1001
time step continuity errors : sum local = 0.0900662, global = -9.51614e-17, cumulative = -8.82929e-17
DICPCG: Solving for p, Initial residual = 0.979527, Final residual = 0.0190689, No Iterations 1001
time step continuity errors : sum local = 2.48803, global = 7.10808e-14, cumulative = 7.09925e-14
ExecutionTime = 0.66 s ClockTime = 0 s
Time = 0.0875
Courant Number mean: 858.698 max: 2.36043e+07
DILUPBiCG: Solving for Ux, Initial residual = 0.999976, Final residual = 1.00727, No Iterations 1001
DILUPBiCG: Solving for Uy, Initial residual = 0.999974, Final residual = 3.69304, No Iterations 1001
DICPCG: Solving for p, Initial residual = 0.999999, Final residual = 0.119223, No Iterations 1001
time step continuity errors : sum local = 2.24302e+07, global = -9.7959e-10, cumulative = -9.79519e-10
DICPCG: Solving for p, Initial residual = 0.981905, Final residual = 0.335337, No Iterations 1001
time step continuity errors : sum local = 3.96954e+09, global = 1.92904e-07, cumulative = 1.91924e-07
ExecutionTime = 0.91 s ClockTime = 1 s
Time = 0.1
Courant Number mean: 8.4947e+09 max: 1.06993e+14
DILUPBiCG: Solving for Ux, Initial residual = 0.997632, Final residual = 51.6564, No Iterations 1001
DILUPBiCG: Solving for Uy, Initial residual = 0.997067, Final residual = 1.91243, No Iterations 1001
DICPCG: Solving for p, Initial residual = 1, Final residual = 12.3467, No Iterations 1001
time step continuity errors : sum local = 3.52888e+15, global = -0.0227559, cumulative = -0.0227557
DICPCG: Solving for p, Initial residual = 0.977625, Final residual = 6.17759, No Iterations 1001
time step continuity errors : sum local = 1.0004e+18, global = 41.4258, cumulative = 41.403
ExecutionTime = 1.17 s ClockTime = 1 s
BTW: I can assure that the settings in OpenFOAM 4.1.0 are correct, because I succeeded in the simulation of one type of structures and have got reasonable results. The meshCheck is also OK.
I have been using OpenFOAM for several months and this is the first time I met this kind of problem.
Any suggestion is appreciated. Thank you in advance.
Best Regards,
Bill
Relevant answer
Answer
Dear Wang,
I got your message and read through your topic here. First of all I agree with all the things mentioned here and would thank Robert Büssing mentioning my book in such a nice way.
I have just a few remarks here:
a) Why do you use icoFoam? This solver does not have an automatic time step adjustment but as you already said, reducing the time step will lead to the same problem; however I would prefer pimpleFoam but of course you can also use icoFoam.
b) A good method to check out the problem regions is to save each time step and check the results in ParaView. Here you should see where the error accumulation starts and you would be able to determine if it is a problem of you mesh (particularly a really bad cell) or if it might be based on numerical discretization. You did not provide any information about the settings you use :) - maybe you are using Gauss linear which is one of the worst discretization schemes for the divergence for example (see also: http://www.holzmann-cfd.de/index.php/en/numerical-schemes)
c) If you do not use Gauss upwind everywhere, test it first. This scheme is highly diffusive and removes all problems (like stiff gradients etc.).
d) Keeping the Reynolds number constant while reducing the velocity field and keeping the geometry the same would lead to an decrease of the viscosity. For me there is no other possible way to achieve the same Reynolds number while reducing the velocity. But I might be wrong here.
e) I prefer using hexadominant cells. As you said, you are using Salome, I guess you are using tetraeder, right? In combination with Gauss linear this is really not recommended as you can see here (https://www.youtube.com/watch?v=C0CcN7l37Fg).
f) If you want, you can send me your *hdf file, if you do not import any file, it is easier to sent me the python script (-> File -> Dump Study).
Good luck.
Tobias Holzmann
  • asked a question related to Iterative Methods
Question
1 answer
Simulink cannot solve the algebraic loop containing 'file name/fractional order derivative' at time 0.1 using the TrustRegion-based algorithm due to one of the following reasons: the model is ill-defined i.e., the system equations do not have a solution; or the nonlinear equation solver failed to converge due to numerical issues.
To rule out solver convergence as the cause of this error
Relevant answer
Answer
Hi, is your problem related to how encapsulate a solver in SIMULINK using S-functions or is the problem related to the non convergence of the Trsut Region method ?
  • asked a question related to Iterative Methods
Question
5 answers
What are the applications of implicit iterative algorithms? Moreover, is there any nonlinear nonexpansive mapping example satisfying some result (any research paper in literature with implicit iteration) having implicit iteration process?
Relevant answer
Answer
I guess Newton's method for optimization qualifies: it is based on the solution of a system of linear equations, and a line search, both of which in general have multiple solutions. Even the very basic algorithm called "steepest descent" is an implicit such method, as it relies on a line search which may have more than one solution. In effect, most known methodologies within nonlinear optimization - at least - are implicit.
  • asked a question related to Iterative Methods
Question
9 answers
I am working on a CFD analysis of a 80m2 greenhouse used for drying seeds. The space has 8 shelves stacked at 4 different heights and openable windows for excess moisture and heat to be removed. I use both '2-k-e' and 'constant effective velocity' solvers and used them separately as well as in succession to generate results. I vary Relaxation Factor on x/y/z velocities, mass and temperature and I also vary Termination Residuals to direct the solver towards convergence.
My question is to anybody with experience of none-symmetrical CFD work at scales close to room/building level; what is a reasonable number of iterations before ditching the simulation as non-convergent? I have had convergence at up to 100,000 iterations when my PC was running over the weekend, so my inclination is to wait, but this particular CFD task leads to a lot of unstable, osculating patterns on Normalised Residuals (see attached) and I wonder if it is worth giving up as soon as the residuals begins to follow similar patterns and ‘repeat’ themselves?
Relevant answer
Answer
Hello,
You can have good guidelines on cfd-online about convergence issues. Meanwhile, you can 
1. Decrease time step (you can adjust the number of iterations as per your requirement of the simulation time)
2. Increase inner iterations
3. Refine mesh appropriately 
4. If you can deal your problem in non-dimensional numbers ( ex. Re ) , using scaled down geometry does improve the convergence.
Hope it helps !
  • asked a question related to Iterative Methods
Question
5 answers
Choosing the initial values of a parameter at beginning of procedure like estimation of parameters is crucial. AS far as I know, good starting point leads to fast converging, for example, the good starting point,  the fast converging in a calculation of hessian matrix.
now, How do I choose an initial value for the parameter?
thanks
Relevant answer
Answer
I would say that scaling your data is the more critical issue.  If you data contains a lot of very small and/or very large numbers, it will be difficult to choose start values.  In fact convergence may fail altogether.  If your data are properly scaled then starting values are easy.  You could then start with a vector of ones, for example. 
If your model is some variation of one with a closed form solution, regression say, you could run a regression for start values. 
  • asked a question related to Iterative Methods
Question
7 answers
Let be [D] a (nxn) ill-conditioned matrix and [G] its inverse. If the inverse exists then it can be efficiently implemented by the linear solving of [D]{gj}={bj} with the LU method using partial pivoting where {gj}: is a column vector of [G] and {bj}: is a column of the identity matrix [I] for j=1 to n and (bj(i)=0 for i=1 to n such that bj(j)=1). The resulting matrix can be checked with the satisfying condition [D][G]=[I].
Relevant answer
Answer
Hi, I am not directly involved in this topic but I think preconditioners may also help:
  • asked a question related to Iterative Methods
Question
9 answers
I was just trying to solve system of transcendental equations using 'Bisection Method'. Iterative procedure is used to check the convergence. In lower order modes, I am unable to obtain the solutions whereas the same code is able to give solutions for arbitrary higher order modes? In those lower order modes, I think, solutions are stuck in LOCAL minima. How can I overcome local minima in 'Iterative Procedure'
Relevant answer
Answer
Respected Prof. Leigh, I've checked these already. In case of 2D or 3D problem, I can check graphically. In higher order problem, 'graphical checking procedure' may not be possible. In that case, how can I check discontinuity?
In code, I can use if-else statement to check the argument. I will do it. Is there any other procedure? I'm waiting for your kind reply, Prof. Leigh.
  • asked a question related to Iterative Methods
Question
3 answers
if we use iteration methods like successive over relaxation, we should use a suitable value of the weight factor and this value will be updated through iteration to force iteration variables to get rapid convergence. my question is how to update this value? what is mathematics we use to do that?
Relevant answer
Answer
Find the relationship between the parameters and use this equation to uddate the factor.
  • asked a question related to Iterative Methods
Question
3 answers
Recently, I obtain a linear system, $Ax = b$, where $A$ is a nonsingular, strictly diagonally dominant $M$-matrix. Then I also got a matrix splitting $A = S - T$, where $S$ is also a nonsingular, strictly diagonally dominant $M$-matrix. So I establish a stationary iteration scheme as follows,
$$Sx^{(k+1)} = Tx^{(k)} + b.$$
According to numerical results, it seems that this iterative scheme is always convergent. Is it rational from theoretical analysis? So, can we prove the convergence of this iterative scheme? i.e.,
show $\rho(S^{-1}T) < 1$, where $\rho(\cdot)$ is the spectral radius.
Please also refer to the URL for details.
Relevant answer
Answer
Zhu is right: the most classical reference is the book by Varga.
Under the conditions you described the method is NOT always convergent: take S=\eps A for constructing a counterexample.
However you can find mild conditions for convergence.. 
In this context there are many beautiful results based on 'monotonicity' arguments
  • asked a question related to Iterative Methods
Question
1 answer
Hi all,
In point (Gauss-Seidel) method of iterative solution, simple algorithm is applied at each point where the function value has to be determined in successive manner, which essentially means that all point methods are a explicit methods. In fact Gauss Seidel method is explicit.
So, How point explicit G-S works ?
whereas in line method of iteration (called implicit method), a row of many grids are solved simultaneously with the known values (existing values) from neighbouring grids. So this method becomes implicit method?
I need a clarification on this form an numerical simulation expert. Please someone answer.
Waiting for a reply.
Thanks
Vivek 
Relevant answer
Answer
In some sense (as e.g. in ode-solution, where a method which requires a linear or nonlinear system solve for one time step is considered as (semi)implicit) the block method is implicit, but in fact, considering a ''small'' linear system solve as trivial, both methods have the same quality:
 for i=1,...,n
 solve a_{ii}x_{i,new} = b_i
     - sum_{j=1 to i-1} a_{i,j} x_{j,new} - sum_{j=i+1 to n}a_{i,j}x_{j,old}
with x_i, a_{ij} scalar  this is the pointwise method and the ''solve'' trivial but you may read this also with x_i of dimension m and the a_{i,j} m*m matrices and hence the whole system of dimension m*n, requiring an exact linear system solve.
In this context the block method does not help much with speed, but in the context of multigrid there are situations where you must use the block method.
  • asked a question related to Iterative Methods
Question
3 answers
I am student of MS Mathematics . currently I am working on this topic  (Nonlinear Complementarity Problem) want to know numerical solution of  such algorithms those have been done on Numeric iterative method .
  • asked a question related to Iterative Methods
Question
17 answers
Consider the problem of fitting a curve of the form
y(x)=a0+a1*exp(-x/b1)+a2*exp(-x/b2)+...+an*exp(-x/bn)
to noisy (x,y) data.  If we have a good initial guess for a0,a1,b1,a2,b2,..., then an iterative method can be used to find a local minimum of the least squares fit to the data.  But is there a good way to generate a good initial guess automatically?
A routine for this purpose, exp2fit.m (see attachment), was contributed to the Matlab File Exchange in 2008 by someone named Per Lundqvist.  It contains some unexplained formulas for generating initial guesses for n=1 and n=2 exponential terms, based on integrals of the data.
-Can anyone explain these formulas?
-Can the approach be extended to the sum of 3 exponentials (and beyond)?
-Are there better alternatives?
Relevant answer
Answer
I just noticed this discussion.
You can consider the so-called Padé-Laplave method (within which a series of other methods can be reformulated as particular, or sub-optimal cases, such as Prony, the method of moments etc). The basic reference is:
Analysis of multiexponential functions without a hypothesis as to the number of components  E. YERAMIAN* & P. CLAVERIE† Nature 326, 169 - 174 (12 March 1987)
regards
  • asked a question related to Iterative Methods
Question
4 answers
If there is no extranous fixed points for any iterative methods, what does mean?
Such methods, possible to diverge at any points in complex plane. ?
If there is many extranous fixed points, then what it means?
For example, one optimal fourth order modified newton's method gives 30 extranous fixed points for solving p(z)=z^3-1. for this results what say's.
Your valuable answer and comments  lead me to good work
Thanks in advance
Relevant answer
Answer
I second Peter's answer: it is often rather easy to transform a problem into a fixed-point one, BUT the properties of the mapping F in the fixed-point problem "F(x) = x" may have weird properties - in particular it may be far from contractive. So I would most often NOT force the original problem to be a fixed-point problem, but to treat it more like what it is when you first formulated in, be it a system of equations, a variational inequality, or indeed an optimization problem with or without constraints.
  • asked a question related to Iterative Methods
Question
3 answers
The method should not be based on optimization technique which takes long time for convergence. Thanks in advance.
Relevant answer
Answer
Hi,
a non-incremental method would have to cope with the rugged nature of the P-V-characteristics. Hence, any method would need a technique to escape local optima, at least without access to additional technical equipment for better measurements. Nevertheless, several heuristics have already been proposed to tackle the problem. If you do not want to use a heuristic, you might be interested in
  • asked a question related to Iterative Methods
Question
10 answers
Hello everybody,
I have implemented a numerical model which consists--among others--of a Finite Difference scheme that is used to solve the quasi-stationary magnetic diffusion equation in three dimensions; i.e., a modified version of the Laplacian operator in Cartesian coordinates. For solving the generated system of linear equations (A)x = b I use the SOR algorithm. The model ran well so far, no matter what input I chose. I should mention that the system matrix A consists of complex-valued elements.
However, for some input parameters the SOR solver does not deliver useful results, or more specifically, it does not converge to the given level of accuracy at all. The system matrix has 8e6x8e6 entries which shouldn't be a problem because, for most input parameters, the SOR algorithm delivers physically correct results. It has to have something to do with the entries of the system matrix per se.
For solving the equation on the Cartesian grid I use the 7-point-stencil. If the unmodified version of the Laplacian operator is solved, the weighing factor at the actual node (i,j,k) is -6. In my modified version, it is -6 + i*x, where x denotes the imaginary part. I even tried SPQR from the SuiteSparse package provided by Tim Davis but since this is not an iterative method MATLAB throws an "out of memory" error (I can use up to 192 GiB (!)). It would really be great if someone encountered a similar problem and wants to share their solution. If interested, I could upload the system matrix as well as the boundary condition vector.
Thank you very much in anticipation!
Best regards,
Chris Volkmar
Relevant answer
Answer
Hi Chris, 
I had the same problem one year ago. Just follow the following steps:
1- Decompose the original system Ax=b into real and imaginary parts, i.e., A=a+jb, then the new system matrix becomes; Bx=b
See O. Z. Mehdizadeh and M. Paraschivoiu, “Investigation of a two-dimensional spectral element method for Helmholtz’s equation,” Journal of Computational Physics, vol. 189, pp. 111-129, 2003.
2- in order to make sure that the system is positive definite, multiply by transpose of B.
3- use the iterative method you like.
4- compose the solution into real and imaginary.
5- Enjoy your solution.
  • asked a question related to Iterative Methods
Question
2 answers
.
Relevant answer
Answer
Have you tried Matlab (iradon()) or Mathematica?
Give a look to the attached paper
Best wishes,
Gianni 
  • asked a question related to Iterative Methods
Question
5 answers
I found some example about iterative learning control but they are not enough for understanding it.
Relevant answer
Answer
Thank for your shares.I want to keep in touch with you.can you send me your works and simulink code.I have found a code about ILC and if you want I can send this code and my works.
  • asked a question related to Iterative Methods
Question
15 answers
Currently I am using Gauss Siedel Iterative method to solve CFD equations. I am facing instability problem i.e. iterations keep on increasing without change in rms value for a given variable such as temperature. 
I want to know which are some stable iterative methods so that one can implement to get rid of diagonal dominance criteria to be satisfied?
Relevant answer
Answer
Dear Paresh, 
From what equations does your matrix come from? Have you looked at the eigenvalues of your matrix?  I ask you these questions because Gauss-Seidel (and Jacboi) only work for symmetric positive definite (SPD) matrices.  This is also true for conjugate gradient (CG). In fact, most theories on convergence of iterative methods are based on SPD matrices.  Boundary conditions may break symmetry for you so this leaves you with moving to GMRES  or BiCGStab.  GMRES  is nothing more than the generalization of CG (where instead of using Lanczos iteration, the Arnoldi iteration is used which is outlined in texts such as Trefethen and Bau).  BiCGStab will also work but it requires two applications of the system matrix (instead of one for GMRES) so we have never seen BiCGStab beat GMRES in efficiency.  I wouldn't worry about preconditioning just now - once you get this to work, then you can worry about efficiency and I would be happy to propose preconditioners to you at that point. 
The first thing you need to do is diagnose why Gauss-Seidel is not working. Can you run a small version of your problem so that you can actually store a small version of your system matrix? If so, then analyze it with, say, Matlab, and get the eigenvalues of your system.  If they are all real and positive, and symmetric then Gauss-Seidel, Jacobi,and CG should work. If they are non-symmetric but real and positive then GMRES will work like a charm. If they are not real and positive, then you will need preconditioning for sure and I can show you how to do this using simple techniques such as polynomial preconditioning.  hope this helps.
  • asked a question related to Iterative Methods
Question
7 answers
I am interested in the numerical solution of convection-diffusion problems where the convection dominates. In the iterative solution with Gauss-Seidel, instabilities can occur for large Peclet numbers. Is the "downwind numbering" as in the paper by Bey & Wittum a possible solution for this problem on structured grids and with variable coefficients (advection speeds)? Unfortunately, I don't have access to the paper.
Relevant answer
Answer
  • asked a question related to Iterative Methods
Question
2 answers
I want to know about the H-optimality. I have found in " A duality theorem and an iterative method of finding h-optimal experimental designs" by V. Melas and S.M.Ermakov.
Relevant answer
Answer
Dear Samrat,
In the world of computation and numerical analysis, theory tells you that the truncation error is always proportional to: h^n, where "h"  is the size of your element (Finite element) or the mesh size (Finite difference) and "n" is the order of approximation; e.g. n = 2 means quadratic approximation and "n=1" means linear approximation.
This fact triggers in your mind the idea that, the more your refine your mesh (reduce h), the lower would be your truncation error and the more accurate is your solution. When you come to apply this concept on your machine (computer), you will find that in often times, if you keep refining, errors will start to grow! and thus too much refinement is equivalent to no refinement (it makes things worse). What cause this trend are the "ROUND-OFF" errors that exist in every machine. So now, almost every computational problem is governed by two sources of errors: Truncation AND Round-off; the total error being their sum.
Eventually, your ultimate goal is to minimize this total Error: Round-Off + Truncation and this can be achieved only be running your simulation at the "OPTIMAL" h.
As far as I know, no one has control over the round-off error dependency over "h", thus you are left with trial simulations to figure out this optimum value.
Thus the H-optimality is all about the procedure to figuring out the optimum mesh size or element size so as to produce minimum total error.  
Hope I answered your question 
  • asked a question related to Iterative Methods
Question
5 answers
I'm doing a relocation using hypoDD software. When I use LSQR for iteration method, everything works fine. But then I try to use SVD to enhance the quality from subset of events and the process stopped.
I have attached the command window when I run hypoDD using SVD. According to the error command, I have try to increase MAXEVE0 in hypoDD.inc into a considerable amount (about a millions).
I also try to remove/purge hypoDD in .bin, and changing MAXEVE0 parameter before reinstalling hypoDD. Those attempts still fail to solve the issue.
Can anyone give me some advices?
Relevant answer
Answer
THERE IS ONE SIMPLE WAY TO OMMIT SUCH DIFFICULT SITUATION.
JUST TRY MATLAB IF IT WOULD BE ABLE TO SOLVE THE PROBLEM BECAUSE THE NUMBER OF ROW IS RATHER BIG.
  • asked a question related to Iterative Methods
Question
9 answers
I want to tune natural frequencies of first few modes of vibration of a shell. Which are the best iterative methods for finite element  model updating of above mentioned problem.
Relevant answer
Answer
For basic knowledges in this topic have a look to this book:
1. Michael Friswell, J.E. Mottershead, "Finite Element Model Updating in Structural Dynamics", Springer Science & Business Media, 1995
These two papers are very good too:
2. E. Dascotte, Model Updating for Structural Dynamics: Past, Present and Future Outlook. Presented at International Conference on Engineering Dynamics (ICED), April 16-18, 2007, Carvoeiro, Algarve, Portugal.
3. John E. Mottersheada, Michael Linkb, Michael I. Friswell, "The sensitivity method in finite element model updating: A tutorial", Mechanical Systems and Signal Processing
Volume 25, Issue 7, October 2011, Pages 2275–2296
  • asked a question related to Iterative Methods
Question
22 answers
Any suggestions?
Relevant answer
Answer
It highly depends on the problem which method will be better. And then you also need to define what you actually mean by "better": Better convergence? Better speed? Sometimes a slightly better convergence is not important because of bad performance. Currently, in our simulation software we are using BiCGSTAB(2). But, we have several problems which we cannot simulate because the solver will not converge. We will try AMG methods for these kinds of problems. But probably it will be too slow in the general case.
  • asked a question related to Iterative Methods
Question
27 answers
Is there a simple condition for global convergence (i.e. regardless of initial guess) of the Newton-Raphson method to solve f(x) = 0, in both the scalar and vector cases? Particularly in the case in which f is not doubly differentiable?
Relevant answer
Answer
Hi, I suggest that you have a look at  the book
Peter Deuflhard, Newton Methods for Nonlinear Problems. Springer, Berlin, 2011.
There is quite a large body of results on global Newton methods (steepest descent, trust region, Levenberg-Marquardt, Newton with damping).
 
  • asked a question related to Iterative Methods
Question
5 answers
I have experience on iterative methods like Jacobi, SOR and AOR and preconditioning as well.
Relevant answer
Peyman, I have parallelized and used the BiCGStab(2) method. It is an iterative method and has been used for solving large and sparse linear equation systems. I suggest you to read one of my papers: "Parallelization of an Iterative Method for Solving Large and Sparse Linear Systems Using the CUDA-Matlab Integration".
  • asked a question related to Iterative Methods
Question
3 answers
Using some iteration method, we can solve a nonlinear system of equation. I know lots of old methods like, Newton Raphson, Gauss newton, marquardt, levenberg marquardt etc. I want to know about some upgraded methods.
Relevant answer
Answer
The methodology of choice depends on several factors: whether the system is described by a differentiable function or not, and in the case that it is whether it has some nice features like diagonal dominance. Just "out of the hat" I give you the following link to a rather recent paper which has several citations to other recent papers and, more importantly, discusses some of the recent add-ons to traditional methods - like the "filter", originating in the extension of an SQP method:
I cannot vouch for the quality of the paper as such but it does have some good and not too old references to start digging further from. Also, it discusses an important aspect from a applications point of view: perhaps not all equations need to be fulfilled exactly while others must be (as they describe, for example, some physical/logical relationships), meaning that the original nonlinear equation can be rewritten as a perhaps slightly easier to solve nonlinear optimization problem with (fewer) constraints.
  • asked a question related to Iterative Methods
Question
6 answers
I have a multiphysics problem, where I have a coupled, PDEs, and and one of them needs the viscosity to be updated. The updating formula uses both the concentration of a substance (from the transport equation), and the velocity (this comes from the other equation. The problem is, the velocity needs the viscosity as well, hence the recursivity. Of course, one can set initial values, but then, I would like to do something similar as in the imperative languages (Fortran, C), where you can set
set initial viscosity
FOR time =1 to N, step deltat
solve flow equation (with a given viscosity for the fluid)
solve transport equation
update viscosity with formula
viscosity(t+1)= viscosity (t) + concentration/velocity
NEXT time step
Can it be done without resorting to MATLAB? if so, how?
I tried on MATLAB, but I couldn't figure out how to call the model, it kept telling me that it could not be found. I do not know if it was that I didn't set a proper path, or was it the loop or something more subtle about the way this recursive functionality must be set up.
Relevant answer
Answer
1. Use Multiphysics feature of Comsol and solve both equations simultaneously
2. Any time you can use Variables to perform computations
3. Read about the prev(.) function
4. Any time you can switch to MATLAB
  • asked a question related to Iterative Methods
Question
4 answers
I have been using ARPACK to calculate a selection of eigenvalues of large-scale problems. Since I am interested in the left- and right eigenvectors I always have to solve the system twice, once for A and once for A^T. The problem is now that the eigenvalues/eigenvectors (left and right) have to be matched and this always leads to problems.
Now, I know that ARPACK is internally using the Arnoldi algorithm which is not designed to calculate the left-eigenvectors simultaniously. However, the only method that I know that does this job is the non-hermitian Lanczos. Unfortunately I could not find any public codes like ARPACK available implementing this code. My questions would be:
1. Is there a software similar to ARPACK that implements a sophisticated version of the non-hermitian Lanczos? If so, can the Matrix-vector product be supplied by myself and can it be connected to an existing C++ code?
2. If I want to implement a sophisticated version of the non-hermitian Lanczos myself, can you tell me some papers that provide implementations coping with the known problems (breakdown, block, etc.).
Relevant answer
Answer
I don't know about the non-hermitian Lanczos, but if ARPACK can do it as described above, you should look at arpack++ http://www.ime.unicamp.br/~chico/arpack++/
and patch http://reuter.mit.edu/software/arpackpatch/ . Arpack++ is a c++ wrapper around arpack that connects to external libraries for the matrix decomposition and product etc.
  • asked a question related to Iterative Methods
Question
18 answers
I have an unknown population distribution, where I want to make inferences of some parameters x that characterize the population distribution. I perform a statistical inference method (e.g Monte Carlo simulations or resampling). The theorem of the central limit says that if I do many repetitions, that is many repeated samples, and I draw the sampling distributions of the means <xi> this should be a normal distribution and in that case my population is large enough.
My sampling distribution of the means <xi> is skewed. Should I perform more simulations to look for the normal distribution? The thing is that I have performed a lot of them, more than 5000 (I think more than sufficient), which makes me think that biases have taken place and hence any parameter value of interest for the population is not representative of the population. Is this correct?
What are the derived conclusions?
Relevant answer
Answer
Central Limit Theorem (CLT) does not work for arbitrarily distributed random numbers. They have to be identically distributed (iid) AND their variances should be finite to make CLT work. For example, the Cauchy distribution, with pdf ~ 1/(x^2 + a^2), or the Pareto-Levy distribution mentioned earlier, are good examples not satisfying those conditions and invalidating CLT.