Mathematical Optimization - History

History

Fermat and Lagrange found calculus-based formulas for identifying optima, while Newton and Gauss proposed iterative methods for moving towards an optimum. Historically, the first term for optimization was "linear programming", which was due to George B. Dantzig, although much of the theory had been introduced by Leonid Kantorovich in 1939. Dantzig published the Simplex algorithm in 1947, and John von Neumann developed the theory of duality in the same year.

The term programming in this context does not refer to computer programming. Rather, the term comes from the use of program by the United States military to refer to proposed training and logistics schedules, which were the problems Dantzig studied at that time.

Later important researchers in mathematical optimization include the following:

  • Richard Bellman
  • Ronald A. Howard
  • Narendra Karmarkar
  • William Karush
  • Leonid Khachiyan
  • Bernard Koopman
  • Harold Kuhn
  • Joseph Louis Lagrange
  • László Lovász
  • Arkadi Nemirovski
  • Yurii Nesterov
  • Boris Polyak
  • Lev Pontryagin
  • James Renegar
  • R. Tyrrell Rockafellar
  • Cornelis Roos
  • Naum Z. Shor
  • Michael J. Todd
  • Albert Tucker

Read more about this topic:  Mathematical Optimization

Famous quotes containing the word history:

    Anything in history or nature that can be described as changing steadily can be seen as heading toward catastrophe.
    Susan Sontag (b. 1933)

    The history of the past is but one long struggle upward to equality.
    Elizabeth Cady Stanton (1815–1902)

    It takes a great deal of history to produce a little literature.
    Henry James (1843–1916)