Search Results for ""
31 - 40 of 1788 for Finite_difference_methodSearch Results
The Wiener-Hopf method is a powerful technique which enables certain linear partial differential equations subject to boundary conditions on semi-infinite domains to be ...
The conjugate gradient method can be viewed as a special variant of the Lanczos method for positive definite symmetric systems. The minimal residual method and symmetric LQ ...
The conjugate gradient method can be viewed as a special variant of the Lanczos method for positive definite symmetric systems. The minimal residual method (MINRES) and ...
Stationary iterative methods are methods for solving a linear system of equations Ax=b, where A is a given matrix and b is a given vector. Stationary iterative methods can be ...
The forward difference is a finite difference defined by Deltaa_n=a_(n+1)-a_n. (1) Higher order differences are obtained by repeated operations of the forward difference ...
A root-finding algorithm based on the iteration formula x_(n+1)=x_n-(f(x_n))/(f^'(x_n)){1+(f(x_n)f^('')(x_n))/(2[f^'(x_n)]^2)}. This method, like Newton's method, has poor ...
The reciprocal differences are closely related to the divided difference. The first few are explicitly given by rho(x_0,x_1)=(x_0-x_1)/(f_0-f_1) (1) ...
A method used by Gauss to solve the quadratic Diophantine equation of the form mx^2+ny^2=A (Dickson 2005, pp. 391 and 407).
The successive square method is an algorithm to compute a^b in a finite field GF(p). The first step is to decompose b in successive powers of two, b=sum_(i)delta_i2^i, (1) ...
Levenberg-Marquardt is a popular alternative to the Gauss-Newton method of finding the minimum of a function F(x) that is a sum of squares of nonlinear functions, ...
...
View search results from all Wolfram sites (32028 matches)

