2010-07-26 Ralf Wildenhues * doc/blas.texi, doc/bspline.texi, doc/complex.texi, doc/dwt.texi, doc/fftalgorithms.tex, doc/fft.texi, doc/fitting.texi, doc/gsl-design.texi, doc/gsl-ref.texi, doc/histogram.texi, doc/integration.texi, doc/linalg.texi, doc/montecarlo.texi, doc/multifit.texi, doc/multimin.texi, doc/ntuple.texi, doc/randist.texi, doc/statnotes.tex: Fix typos. diff -ru orig/gsl-1.14/doc/blas.texi gsl-1.14/doc/blas.texi --- orig/gsl-1.14/doc/blas.texi 2010-03-10 11:57:12.000000000 +0100 +++ gsl-1.14/doc/blas.texi 2010-07-26 14:41:50.000000000 +0200 @@ -22,7 +22,7 @@ functions. The full @sc{blas} functionality for band-format and packed-format matrices is available through the low-level @sc{cblas} interface. Similarly, GSL vectors are restricted to positive strides, -whereas the the low-level @sc{cblas} interface supports negative +whereas the low-level @sc{cblas} interface supports negative strides as specified in the @sc{blas} address@hidden the low-level @sc{cblas} interface, a negative stride accesses the vector elements in reverse order, i.e. the @math{i}-th element is given by diff -ru orig/gsl-1.14/doc/bspline.texi gsl-1.14/doc/bspline.texi --- orig/gsl-1.14/doc/bspline.texi 2010-03-10 11:57:12.000000000 +0100 +++ gsl-1.14/doc/bspline.texi 2010-07-26 14:46:25.000000000 +0200 @@ -227,7 +227,7 @@ The following program computes a linear least squares fit to data using cubic B-spline basis functions with uniform breakpoints. The data is generated from the curve @math{y(x) = \cos{(x)} \exp{(-x/10)}} on -the interval @math{[0, 15]} with gaussian noise added. +the interval @math{[0, 15]} with Gaussian noise added. @example @verbatiminclude examples/bspline.c diff -ru orig/gsl-1.14/doc/complex.texi gsl-1.14/doc/complex.texi --- orig/gsl-1.14/doc/complex.texi 2010-03-10 11:57:12.000000000 +0100 +++ gsl-1.14/doc/complex.texi 2010-07-26 14:56:25.000000000 +0200 @@ -60,7 +60,7 @@ be mapped correctly onto packed complex arrays. @deftypefun gsl_complex gsl_complex_rect (double @var{x}, double @var{y}) -This function uses the rectangular cartesian components +This function uses the rectangular Cartesian components (@var{x},@var{y}) to return the complex number @math{z = x + i y}. @inlinefn{} @end deftypefun @@ -77,7 +77,7 @@ @end defmac @defmac GSL_SET_COMPLEX (@var{zp}, @var{x}, @var{y}) -This macro uses the cartesian components (@var{x},@var{y}) to set the +This macro uses the Cartesian components (@var{x},@var{y}) to set the real and imaginary parts of the complex number pointed to by @var{zp}. For example, diff -ru orig/gsl-1.14/doc/dwt.texi gsl-1.14/doc/dwt.texi --- orig/gsl-1.14/doc/dwt.texi 2010-03-10 11:57:12.000000000 +0100 +++ gsl-1.14/doc/dwt.texi 2010-07-26 15:00:13.000000000 +0200 @@ -276,7 +276,7 @@ transform on the rows of the matrix, followed by a separate complete discrete wavelet transform on the columns of the resulting row-transformed matrix. This procedure uses the same ordering as a -two-dimensional fourier transform. +two-dimensional Fourier transform. The ``non-standard'' transform is performed in interleaved passes on the rows and columns of the matrix for each level of the transform. The diff -ru orig/gsl-1.14/doc/fftalgorithms.tex gsl-1.14/doc/fftalgorithms.tex --- orig/gsl-1.14/doc/fftalgorithms.tex 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/fftalgorithms.tex 2010-07-26 14:14:53.000000000 +0200 @@ -21,7 +21,7 @@ \section{Introduction} Fast Fourier Transforms (FFTs) are efficient algorithms for -calculating the discrete fourier transform (DFT), +calculating the discrete Fourier transform (DFT), % \begin{eqnarray} h_a &=& \mathrm{DFT}(g_b) \\ @@ -29,9 +29,9 @@ &=& \sum_{b=0}^{N-1} g_b W_N^{ab} \qquad W_N= \exp(-2\pi i/N) \end{eqnarray} % -The DFT usually arises as an approximation to the continuous fourier +The DFT usually arises as an approximation to the continuous Fourier transform when functions are sampled at discrete intervals in space or -time. The naive evaluation of the discrete fourier transform is a +time. The naive evaluation of the discrete Fourier transform is a matrix-vector multiplication ${\mathbf W}\vec{g}$, and would take $O(N^2)$ operations for $N$ data-points. The general principle of the Fast Fourier Transform algorithms is to use a divide-and-conquer @@ -334,7 +334,7 @@ \subsection{Radix-2 Decimation-in-Time (DIT)} % -To derive the the decimation-in-time algorithm we start by separating +To derive the decimation-in-time algorithm we start by separating out the most significant bit of the index $b$, % \begin{equation} @@ -504,7 +504,7 @@ So for an in-place pass our storage has to be arranged so that the two outputs $g_1(a_0,\dots)$ overwrite the two input terms $g([b_{n-1},\dots])$. Note that the order of $a$ is reversed from the -natural order of $b$. i.e. the least significant bit of $a$ +natural order of $b$, i.e.@: the least significant bit of $a$ replaces the most significant bit of $b$. This is inconvenient because $a$ occurs in its natural order in all the exponentials, $W^{ab}$. We could keep track of both $a$ and its bit-reverse, @@ -1362,7 +1362,7 @@ $p_{i-1}$ independent multiplications of $PD$ on $q_{i-1}$ different subsets of $t$. The index $\mu$ of $t(\lambda,\mu)$ which runs from 0 to $m$ will include $q_i$ copies of each $PD$ operation because -$m=p_{i-1}q$. i.e. we can split the index $\mu$ further into $\mu = a +$m=p_{i-1}q$, i.e.@: we can split the index $\mu$ further into $\mu = a p_{i-1} + b$, where $a = 0 \dots q-1$ and $b=0 \dots p_{i-1}$, % \begin{eqnarray} @@ -1571,7 +1571,7 @@ $\omega^a_{q_{i-1}}$ are taken out of the {\tt trig} array. To compute the inverse transform we go back to the definition of the -fourier transform and note that the inverse matrix is just the complex +Fourier transform and note that the inverse matrix is just the complex conjugate of the forward matrix (with a factor of $1/N$), % \begin{equation} @@ -1784,7 +1784,7 @@ for computing a DFT~\cite{singleton}. Although it is an $O(N^2)$ algorithm it does reduce the number of multiplications by a factor of 4 compared with a naive evaluation of the DFT. If we look at the -general stucture of a DFT matrix, shown schematically below, +general structure of a DFT matrix, shown schematically below, % \begin{equation} \left( @@ -2490,7 +2490,7 @@ \subsection{Mixed-Radix FFTs for real data} % As discussed earlier the radix-2 decimation-in-time algorithm had the -special property that its intermediate passes are interleaved fourier +special property that its intermediate passes are interleaved Fourier transforms of the original data, and this generalizes to the mixed-radix algorithm. The complex mixed-radix algorithm that we derived earlier was a decimation-in-frequency algorithm, but we can @@ -2580,7 +2580,7 @@ v^{(i)} = (W_{p_i} \otimes I_{q_i}) z \end{equation} % -Each intermediate stage will be a set of $q_i$ interleaved fourier +Each intermediate stage will be a set of $q_i$ interleaved Fourier transforms, each of length $p_i$. We can prove this result by induction. First we assume that the result is true for $v^{(i-1)}$, % @@ -2634,7 +2634,7 @@ explicitly, and induction then shows that the result is true for all $i$. As discussed for the radix-2 algorithm this result is important because if the initial data $z$ is real then each intermediate pass is -a set of interleaved fourier transforms of $z$, having half-complex +a set of interleaved Fourier transforms of $z$, having half-complex symmetries (appropriately applied in the subspaces of the Kronecker product). Consequently only $N$ real numbers are needed to store the intermediate and final results. diff -ru orig/gsl-1.14/doc/fft.texi gsl-1.14/doc/fft.texi --- orig/gsl-1.14/doc/fft.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/fft.texi 2010-07-26 15:06:24.000000000 +0200 @@ -31,7 +31,7 @@ @cindex FFT mathematical definition Fast Fourier Transforms are efficient algorithms for -calculating the discrete fourier transform (DFT), +calculating the discrete Fourier transform (DFT), @tex \beforedisplay $$ @@ -46,13 +46,13 @@ @end example @end ifinfo -The DFT usually arises as an approximation to the continuous fourier +The DFT usually arises as an approximation to the continuous Fourier transform when functions are sampled at discrete intervals in space or -time. The naive evaluation of the discrete fourier transform is a +time. The naive evaluation of the discrete Fourier transform is a matrix-vector multiplication @c{$W\vec{z}$} @address@hidden@}}. A general matrix-vector multiplication takes address@hidden(n^2)} operations for @math{n} data-points. Fast fourier address@hidden(n^2)} operations for @math{n} data-points. Fast Fourier transform algorithms use a divide-and-conquer strategy to factorize the matrix @math{W} into smaller sub-matrices, corresponding to the integer factors of the length @math{n}. If @math{n} can be factorized into a @@ -64,7 +64,7 @@ All the FFT functions offer three types of transform: forwards, inverse and backwards, based on the same mathematical definitions. The -definition of the @dfn{forward fourier transform}, +definition of the @dfn{forward Fourier transform}, @c{$x = \hbox{FFT}(z)$} @math{x = FFT(z)}, is, @tex @@ -82,7 +82,7 @@ @end ifinfo @noindent -and the definition of the @dfn{inverse fourier transform}, +and the definition of the @dfn{inverse Fourier transform}, @c{$x = \hbox{IFFT}(z)$} @math{x = IFFT(z)}, is, @tex @@ -109,7 +109,7 @@ exponential in the transform/ inverse-transform pair. GSL follows the same convention as @sc{fftpack}, using a negative exponential for the forward transform. The advantage of this convention is that the inverse -transform recreates the original function with simple fourier +transform recreates the original function with simple Fourier synthesis. Numerical Recipes uses the opposite convention, a positive exponential in the forward transform. @@ -269,7 +269,7 @@ @comment @subsection Example of using radix-2 FFT routines for complex data Here is an example program which computes the FFT of a short pulse in a -sample of length 128. To make the resulting fourier transform real the +sample of length 128. To make the resulting Fourier transform real the pulse is defined for equal positive and negative times (@math{-10} @dots{} @math{10}), where the negative times wrap around the end of the array. @@ -288,7 +288,7 @@ the same plot as the input. Only the real part is shown, by the choice of the input data the imaginary part is zero. Allowing for the wrap-around of negative times at @math{t=128}, and working in units of address@hidden/n}, the DFT approximates the continuum fourier transform, giving address@hidden/n}, the DFT approximates the continuum Fourier transform, giving a modulated sine function. @iftex @tex @@ -303,7 +303,7 @@ @center @image{fft-complex-radix2-t,2.8in} @center @image{fft-complex-radix2-f,2.8in} @quotation -A pulse and its discrete fourier transform, output from +A pulse and its discrete Fourier transform, output from the example program. @end quotation @end iftex @@ -513,7 +513,7 @@ @cindex FFT of real data The functions for real data are similar to those for complex data. However, there is an important difference between forward and inverse -transforms. The fourier transform of a real sequence is not real. It is +transforms. The Fourier transform of a real sequence is not real. It is a complex sequence with a special symmetry: @tex \beforedisplay @@ -540,7 +540,7 @@ Functions in @code{gsl_fft_real} compute the frequency coefficients of a real sequence. The half-complex coefficients @math{c} of a real sequence address@hidden are given by fourier analysis, address@hidden are given by Fourier analysis, @tex \beforedisplay $$ @@ -557,7 +557,7 @@ @end ifinfo @noindent Functions in @code{gsl_fft_halfcomplex} compute inverse or backwards -transforms. They reconstruct real sequences by fourier synthesis from +transforms. They reconstruct real sequences by Fourier synthesis from their half-complex frequency coefficients, @math{c}, @tex \beforedisplay @@ -832,7 +832,7 @@ array of length @var{n}, using a mixed radix decimation-in-frequency algorithm. For @code{gsl_fft_real_transform} @var{data} is an array of time-ordered real data. For @code{gsl_fft_halfcomplex_transform} address@hidden contains fourier coefficients in the half-complex ordering address@hidden contains Fourier coefficients in the half-complex ordering described above. There is no restriction on the length @var{n}. Efficient modules are provided for subtransforms of length 2, 3, 4 and 5. Any remaining factors are computed with a slow, @math{O(n^2)}, @@ -902,14 +902,14 @@ Here is an example program using @code{gsl_fft_real_transform} and @code{gsl_fft_halfcomplex_inverse}. It generates a real signal in the -shape of a square pulse. The pulse is fourier transformed to frequency +shape of a square pulse. The pulse is Fourier transformed to frequency space, and all but the lowest ten frequency components are removed from -the array of fourier coefficients returned by +the array of Fourier coefficients returned by @code{gsl_fft_real_transform}. -The remaining fourier coefficients are transformed back to the +The remaining Fourier coefficients are transformed back to the time-domain, to give a filtered version of the square pulse. Since -fourier coefficients are stored using the half-complex symmetry both +Fourier coefficients are stored using the half-complex symmetry both positive and negative frequencies are removed and the final filtered signal is also real. @@ -935,7 +935,7 @@ @itemize @w{} @item P. Duhamel and M. Vetterli. -Fast fourier transforms: A tutorial review and a state of the art. +Fast Fourier transforms: A tutorial review and a state of the art. @cite{Signal Processing}, 19:259--299, 1990. @end itemize @@ -972,7 +972,7 @@ @itemize @w{} @item Clive Temperton. -Self-sorting mixed-radix fast fourier transforms. +Self-sorting mixed-radix fast Fourier transforms. @cite{Journal of Computational Physics}, 52(1):1--23, 1983. @end itemize @@ -984,13 +984,13 @@ @item Henrik V. Sorenson, Douglas L. Jones, Michael T. Heideman, and C. Sidney Burrus. -Real-valued fast fourier transform algorithms. +Real-valued fast Fourier transform algorithms. @cite{IEEE Transactions on Acoustics, Speech, and Signal Processing}, ASSP-35(6):849--863, 1987. @item Clive Temperton. -Fast mixed-radix real fourier transforms. +Fast mixed-radix real Fourier transforms. @cite{Journal of Computational Physics}, 52:340--350, 1983. @end itemize diff -ru orig/gsl-1.14/doc/fitting.texi gsl-1.14/doc/fitting.texi --- orig/gsl-1.14/doc/fitting.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/fitting.texi 2010-07-26 15:07:05.000000000 +0200 @@ -50,7 +50,7 @@ weight factors @math{w_i} are given by @math{w_i = 1/\sigma_i^2}, where @math{\sigma_i} is the experimental error on the data-point @math{y_i}. The errors are assumed to be -gaussian and uncorrelated. +Gaussian and uncorrelated. For unweighted data the chi-squared sum is computed without any weight factors. The fitting routines return the best-fit parameters @math{c} and their @@ -60,7 +60,7 @@ @cindex covariance matrix, linear fits as @c{$C_{ab} = \langle \delta c_a \delta c_b \rangle$} @address@hidden@} = <\delta c_a \delta c_b>} where @c{$\langle \, \rangle$} address@hidden< >} denotes an average over the gaussian error distributions of the underlying datapoints. address@hidden< >} denotes an average over the Gaussian error distributions of the underlying datapoints. The covariance matrix is calculated by error propagation from the data errors @math{\sigma_i}. The change in a fitted parameter @math{\delta diff -ru orig/gsl-1.14/doc/gsl-design.texi gsl-1.14/doc/gsl-design.texi --- orig/gsl-1.14/doc/gsl-design.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/gsl-design.texi 2010-07-26 15:13:09.000000000 +0200 @@ -386,7 +386,7 @@ @c reliable and accurate (but not necessarily fast or efficient) estimation @c of values for special functions, explicitly using Taylor series, asymptotic @c expansions, continued fraction expansions, etc. As well as these routines, address@hidden fast approximations will also be provided, primarily based on Chebyschev address@hidden fast approximations will also be provided, primarily based on Chebyshev @c polynomials and ratios of polynomials. In this vision, the approximations @c will be the "standard" routines for the users, and the exact (so-called) @c routines will be used for verification of the approximations. It may also @@ -413,7 +413,7 @@ @c @item Direct integration address@hidden @item Monte carlo methods address@hidden @item Monte Carlo methods @c @item Simulated annealing @@ -459,12 +459,12 @@ "closed". In mathematics objects can be combined and operated on in an infinite number of ways. For example, I can take the derivative of a scalar field with respect to a vector and the derivative of a vector -field wrt a scalar (along a path). +field wrt.@: a scalar (along a path). There is a definite tendency to unconsciously try to reproduce all these possibilities in a numerical library, by adding new features one by one. After all, it is always easy enough to support just one more -feature.... so why not? +feature @dots{} so why not? Looking at the big picture, no-one would start out by saying "I want to be able to represent every possible mathematical object and operation @@ -660,7 +660,7 @@ should be to Knuth, references concerning statistics should be to Kendall & Stuart, references concerning special functions should be to Abramowitz & Stegun (Handbook of Mathematical Functions AMS-55), etc. -Whereever possible refer to Abramowitz & Stegun rather than other +Wherever possible refer to Abramowitz & Stegun rather than other reference books because it is a public domain work, so it is inexpensive and freely redistributable. @@ -711,16 +711,16 @@ and Texinfo. This is a problem if you want to write something like @address@hidden@}}. -To work around it you can preceed the math command with a special +To work around it you can precede the math command with a special macro @code{@@c} which contains the explicit TeX commands you want to use (no restrictions), and put an ASCII approximation into the @code{@@math} command (you can write @code{@@@{} and @code{@@@}} there for the left and right braces). The explicit TeX -commands are used in the TeX ouput and the argument of @code{@@math} +commands are used in the TeX output and the argument of @code{@@math} in the plain info output. Note that the @code{@@address@hidden@}} macro must go at the end of the -preceeding line, because everything else after it is ignored---as far +preceding line, because everything else after it is ignored---as far as texinfo is concerned it's actually a 'comment'. The comment command @@c has been modified to capture a TeX expression which is output by the next @@math command. For ordinary comments use the @@comment @@ -763,7 +763,7 @@ Any installed executables (utility programs etc) should have the prefix @code{gsl-} (with a hyphen, not an underscore). -All function names, variables, etc should be in lower case. Macros and +All function names, variables, etc.@: should be in lower case. Macros and preprocessor variables should be in upper case. Some common conventions in variable and function names: @@ -816,12 +816,12 @@ Note: it is possible to define an abstract base class easily in C, using function pointers. See the rng directory for an example. -When reimplementing public domain fortran code, please try to introduce +When reimplementing public domain Fortran code, please try to introduce the appropriate object concepts as structs, rather than translating the code literally in terms of arrays. The structs can be useful just within the file, you don't need to export them to the user. -For example, if a fortran program repeatedly uses a subroutine like, +For example, if a Fortran program repeatedly uses a subroutine like, @example SUBROUTINE RESIZE (X, K, ND, K1) @@ -954,10 +954,10 @@ @section Error estimates In the special functions error bounds are given as twice the expected -``gaussian'' error. i.e. 2-sigma, so the result is inside the error +``Gaussian'' error, i.e.@: 2-sigma, so the result is inside the error 98% of the time. People expect the true value to be within +/- the quoted error (this wouldn't be the case 32% of the time for 1 sigma). -Obviously the errors are not gaussian but a factor of two works well +Obviously the errors are not Gaussian but a factor of two works well in practice. @node Exceptions and Error handling, Persistence, Error estimates, Design @@ -1293,7 +1293,7 @@ significant or not. The only place where it is acceptable to use constants like address@hidden is in function approximations, (e.g. taylor address@hidden is in function approximations, (e.g.@: Taylor series, asymptotic expansions, etc). In these cases it is not an arbitrary constant, but an inherent part of the algorithm. @@ -1406,7 +1406,7 @@ @smallexample Yoyodyne, Inc., hereby disclaims all copyright interest in the software `GNU Scientific Library - Legendre Functions' (routines for computing -legendre functions numerically in C) written by James Hacker. +Legendre functions numerically in C) written by James Hacker. , 1 April 1989 Ty Coon, President of Vice @@ -1572,7 +1572,7 @@ ISBN 0521483913. @item address@hidden Transcendental Functions satisfying nonhomogenous linear differential equations} by A W Babister, address@hidden Transcendental Functions satisfying nonhomogeneous linear differential equations} by A W Babister, ISBN 1114401773. @end itemize diff -ru orig/gsl-1.14/doc/gsl-ref.texi gsl-1.14/doc/gsl-ref.texi --- orig/gsl-1.14/doc/gsl-ref.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/gsl-ref.texi 2010-07-26 15:14:23.000000000 +0200 @@ -499,7 +499,7 @@ Wrote the initial complex arithmetic functions. @item Thomas Walter -Wrote the initial heapsort routines and cholesky decomposition. +Wrote the initial heapsort routines and Cholesky decomposition. @item Fabrice Rossi Multidimensional minimization. diff -ru orig/gsl-1.14/doc/histogram.texi gsl-1.14/doc/histogram.texi --- orig/gsl-1.14/doc/histogram.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/histogram.texi 2010-07-26 15:24:58.000000000 +0200 @@ -760,7 +760,7 @@ @comment @deftypefun {gsl_histogram2d *} gsl_histogram2d_calloc_range (size_t @var{nx}, size_t @var{ny}, double * @var{xrange}, double * @var{yrange}) @comment This function allocates a histogram of size @address@hidden using @comment the @math{nx+1} and @math{ny+1} bin ranges specified by the arrays address@hidden @var{xrange} and @var{xyrange}. address@hidden @var{xrange} and @var{yrange}. @comment @end deftypefun @deftypefun int gsl_histogram2d_set_ranges (gsl_histogram2d * @var{h}, const double @var{xrange}[], size_t @var{xsize}, const double @var{yrange}[], size_t @var{ysize}) diff -ru orig/gsl-1.14/doc/integration.texi gsl-1.14/doc/integration.texi --- orig/gsl-1.14/doc/integration.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/integration.texi 2010-07-26 15:30:43.000000000 +0200 @@ -11,7 +11,7 @@ logarithmic singularities, computation of Cauchy principal values and oscillatory integrals. The library reimplements the algorithms used in @sc{quadpack}, a numerical integration package written by Piessens, -Doncker-Kapenga, Uberhuber and Kahaner. Fortran code for @sc{quadpack} is +de Doncker-Kapenga, Ueberhuber and Kahaner. Fortran code for @sc{quadpack} is available on Netlib. Also included are non-adaptive, fixed-order Gauss-Legendre integration routines with high precision coefficients by Pavel Holoborodko. @@ -695,7 +695,7 @@ The parameter @math{\omega} and choice of @math{\sin} or @math{\cos} is taken from the table @var{wf} (the length @var{L} can take any value, since it is overridden by this function to a value appropriate for the -fourier integration). The integral is computed using the QAWO algorithm +Fourier integration). The integral is computed using the QAWO algorithm over each of the subintervals, @tex \beforedisplay @@ -884,7 +884,7 @@ @itemize @w{} @item -R. Piessens, E. de Doncker-Kapenga, C.W. Uberhuber, D.K. Kahaner. +R. Piessens, E. de Doncker-Kapenga, C.W. Ueberhuber, D.K. Kahaner. @address@hidden A subroutine package for automatic integration} Springer Verlag, 1983. @end itemize diff -ru orig/gsl-1.14/doc/linalg.texi gsl-1.14/doc/linalg.texi --- orig/gsl-1.14/doc/linalg.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/linalg.texi 2010-07-26 15:42:11.000000000 +0200 @@ -665,7 +665,7 @@ @node Hessenberg Decomposition of Real Matrices @section Hessenberg Decomposition of Real Matrices address@hidden hessenberg decomposition address@hidden Hessenberg decomposition A general real matrix @math{A} can be decomposed by orthogonal similarity transformations into the form @@ -725,7 +725,7 @@ @node Hessenberg-Triangular Decomposition of Real Matrices @section Hessenberg-Triangular Decomposition of Real Matrices address@hidden hessenberg triangular decomposition address@hidden Hessenberg triangular decomposition A general real matrix pair (@math{A}, @math{B}) can be decomposed by orthogonal similarity transformations into the form @@ -1191,7 +1191,7 @@ 268--275. @item -James Demmel, Kresimir Veselic, ``Jacobi's Method is more accurate than +James Demmel, address@hidden Veseli@'c, ``Jacobi's Method is more accurate than QR'', @cite{Lapack Working Note 15} (LAWN-15), October 1989. Available from netlib, @uref{http://www.netlib.org/lapack/} in the @code{lawns} or @code{lawnspdf} directories. diff -ru orig/gsl-1.14/doc/montecarlo.texi gsl-1.14/doc/montecarlo.texi --- orig/gsl-1.14/doc/montecarlo.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/montecarlo.texi 2010-07-26 15:44:26.000000000 +0200 @@ -1,5 +1,5 @@ @cindex Monte Carlo integration address@hidden stratified sampling in monte carlo integration address@hidden stratified sampling in Monte Carlo integration @cindex multidimensional integration This chapter describes routines for multidimensional Monte Carlo integration. These include the traditional Monte Carlo method and @@ -143,7 +143,7 @@ @node PLAIN Monte Carlo @section PLAIN Monte Carlo address@hidden plain monte carlo address@hidden plain Monte Carlo The plain Monte Carlo algorithm samples points randomly from the integration region to estimate the integral and its error. Using this algorithm the estimate of the integral @math{E(f; N)} for @math{N} @@ -410,7 +410,7 @@ @node VEGAS @section VEGAS address@hidden VEGAS monte carlo integration address@hidden VEGAS Monte Carlo integration @cindex importance sampling, VEGAS The @sc{vegas} algorithm of Lepage is based on importance sampling. It @@ -600,7 +600,7 @@ @deftypevar int stage Setting this determines the @dfn{stage} of the calculation. Normally, @code{stage = 0} which begins with a new uniform grid and empty weighted -average. Calling vegas with @code{stage = 1} retains the grid from the +average. Calling @sc{vegas} with @code{stage = 1} retains the grid from the previous run but discards the weighted average, so that one can ``tune'' the grid using a relatively small number of points and then do a large run with @code{stage = 1} on the optimized grid. Setting @code{stage = diff -ru orig/gsl-1.14/doc/multifit.texi gsl-1.14/doc/multifit.texi --- orig/gsl-1.14/doc/multifit.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/multifit.texi 2010-07-26 15:45:37.000000000 +0200 @@ -77,7 +77,7 @@ regime. To perform a weighted least-squares fit of a nonlinear model address@hidden(x,t)} to data (@math{t_i}, @math{y_i}) with independent gaussian address@hidden(x,t)} to data (@math{t_i}, @math{y_i}) with independent Gaussian errors @math{\sigma_i}, use function components of the following form, @tex \beforedisplay @@ -507,7 +507,7 @@ If the minimisation uses the weighted least-squares function @math{f_i = (Y(x, t_i) - y_i) / \sigma_i} then the covariance matrix above gives the statistical error on the best-fit parameters -resulting from the gaussian errors @math{\sigma_i} on +resulting from the Gaussian errors @math{\sigma_i} on the underlying data @math{y_i}. This can be verified from the relation @math{\delta f = J \delta c} and the fact that the fluctuations in @math{f} from the data @math{y_i} are normalised by @math{\sigma_i} and @@ -576,7 +576,7 @@ @noindent The main part of the program sets up a Levenberg-Marquardt solver and some simulated random data. The data uses the known parameters -(1.0,5.0,0.1) combined with gaussian noise (standard deviation = 0.1) +(1.0,5.0,0.1) combined with Gaussian noise (standard deviation = 0.1) over a range of 40 timesteps. The initial guess for the parameters is chosen as (0.0, 1.0, 0.0). @@ -624,7 +624,7 @@ @address@hidden/address@hidden in this case, a common way of increasing the errors for a poor fit. Note that a poor fit will result from the use an inappropriate model, and the scaled error estimates may then -be outside the range of validity for gaussian errors. +be outside the range of validity for Gaussian errors. @iftex @sp 1 diff -ru orig/gsl-1.14/doc/multimin.texi gsl-1.14/doc/multimin.texi --- orig/gsl-1.14/doc/multimin.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/multimin.texi 2010-07-26 15:46:52.000000000 +0200 @@ -402,7 +402,7 @@ The @code{bfgs2} version of this minimizer is the most efficient version available, and is a faithful implementation of the line minimization scheme described in Fletcher's @cite{Practical Methods of -Optimization}, Algorithms 2.6.2 and 2.6.4. It supercedes the original +Optimization}, Algorithms 2.6.2 and 2.6.4. It supersedes the original @code{bfgs} routine and requires substantially fewer function and gradient evaluations. The user-supplied tolerance @var{tol} corresponds to the parameter @math{\sigma} used by Fletcher. A value diff -ru orig/gsl-1.14/doc/ntuple.texi gsl-1.14/doc/ntuple.texi --- orig/gsl-1.14/doc/ntuple.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/ntuple.texi 2010-07-26 15:48:35.000000000 +0200 @@ -159,7 +159,7 @@ The following example programs demonstrate the use of ntuples in managing a large dataset. The first program creates a set of 10,000 simulated ``events'', each with 3 associated values @math{(x,y,z)}. These -are generated from a gaussian distribution with unit variance, for +are generated from a Gaussian distribution with unit variance, for demonstration purposes, and written to the ntuple file @file{test.dat}. @example diff -ru orig/gsl-1.14/doc/randist.texi gsl-1.14/doc/randist.texi --- orig/gsl-1.14/doc/randist.texi 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/randist.texi 2010-07-26 15:54:51.000000000 +0200 @@ -482,7 +482,7 @@ @noindent for @c{$x \ge 0$} @math{x >= 0}. For @math{b = 1} this reduces to the Laplace -distribution. For @math{b = 2} it has the same form as a gaussian +distribution. For @math{b = 2} it has the same form as a Gaussian distribution, but with @c{$a = \sqrt{2} \sigma$} @math{a = address@hidden@} \sigma}. @end deftypefun @@ -694,7 +694,7 @@ @cindex Levy distribution This function returns a random variate from the Levy symmetric stable distribution with scale @var{c} and exponent @var{alpha}. The symmetric -stable probability distribution is defined by a fourier transform, +stable probability distribution is defined by a Fourier transform, @tex \beforedisplay $$ @@ -737,7 +737,7 @@ distribution with scale @var{c}, exponent @var{alpha} and skewness parameter @var{beta}. The skewness parameter must lie in the range @math{[-1,1]}. The Levy skew stable probability distribution is defined -by a fourier transform, +by a Fourier transform, @tex \beforedisplay $$ @@ -944,7 +944,7 @@ @node The Chi-squared Distribution @section The Chi-squared Distribution The chi-squared distribution arises in statistics. If @math{Y_i} are address@hidden independent gaussian random variates with unit variance then the address@hidden independent Gaussian random variates with unit variance then the sum-of-squares, @tex \beforedisplay @@ -1344,8 +1344,8 @@ such that @c{$|v|^2 = x_1^2 + x_2^2 + \cdots + x_n^2 = 1$} @math{|v|^2 = x_1^2 + x_2^2 + ... + x_n^2 = 1}. The method -uses the fact that a multivariate gaussian distribution is spherically -symmetric. Each component is generated to have a gaussian distribution, +uses the fact that a multivariate Gaussian distribution is spherically +symmetric. Each component is generated to have a Gaussian distribution, and then the components are normalized. The method is described by Knuth, v2, 3rd ed, p135--136, and attributed to G. W. Brown, Modern Mathematics for the Engineer (1956). Only in gsl-1.14/doc: .rng.texi.swp Only in gsl-1.14/doc: spell.add Only in gsl-1.14/doc: spell.add1 Only in gsl-1.14/doc: spell.add.spl diff -ru orig/gsl-1.14/doc/statnotes.tex gsl-1.14/doc/statnotes.tex --- orig/gsl-1.14/doc/statnotes.tex 2010-03-10 11:57:13.000000000 +0100 +++ gsl-1.14/doc/statnotes.tex 2010-07-26 14:16:33.000000000 +0200 @@ -7,7 +7,7 @@ \maketitle \section{Weighted mean and variance} -We have $N$ samples $x_i$ drawn from a gaussian distribution +We have $N$ samples $x_i$ drawn from a Gaussian distribution $G(\mu,\sigma)$ (or any distribution with finite first and second moments). Each sample has a weight $w_i$ which represents the relative value we place on it. Given the estimate of the mean @@ -86,7 +86,7 @@ If we assume that the ``experimental errors'' arising from the weights contribute, the underlying variance $\sigma^2$ is overestimated by this formula (e.g. consider the case $\sigma = 0$---all the variation -will come from the gaussian fluctuations represented by the +will come from the Gaussian fluctuations represented by the $w_i$). The appropriate expectation in this case is $\expectation{x_i x_j} = \mu^2 + \delta_{ij} (\sigma^2 + 1/w_i)$ \end{document}