Mathematical Optimization - History

History

Fermat and Lagrange found calculus-based formulas for identifying optima, while Newton and Gauss proposed iterative methods for moving towards an optimum. Historically, the first term for optimization was "linear programming", which was due to George B. Dantzig, although much of the theory had been introduced by Leonid Kantorovich in 1939. Dantzig published the Simplex algorithm in 1947, and John von Neumann developed the theory of duality in the same year.

The term programming in this context does not refer to computer programming. Rather, the term comes from the use of program by the United States military to refer to proposed training and logistics schedules, which were the problems Dantzig studied at that time.

Later important researchers in mathematical optimization include the following:

  • Richard Bellman
  • Ronald A. Howard
  • Narendra Karmarkar
  • William Karush
  • Leonid Khachiyan
  • Bernard Koopman
  • Harold Kuhn
  • Joseph Louis Lagrange
  • László Lovász
  • Arkadi Nemirovski
  • Yurii Nesterov
  • Boris Polyak
  • Lev Pontryagin
  • James Renegar
  • R. Tyrrell Rockafellar
  • Cornelis Roos
  • Naum Z. Shor
  • Michael J. Todd
  • Albert Tucker

Read more about this topic:  Mathematical Optimization

Famous quotes containing the word history:

    To a surprising extent the war-lords in shining armour, the apostles of the martial virtues, tend not to die fighting when the time comes. History is full of ignominious getaways by the great and famous.
    George Orwell (1903–1950)

    No matter how vital experience might be while you lived it, no sooner was it ended and dead than it became as lifeless as the piles of dry dust in a school history book.
    Ellen Glasgow (1874–1945)

    Let it suffice that in the light of these two facts, namely, that the mind is One, and that nature is its correlative, history is to be read and written.
    Ralph Waldo Emerson (1803–1882)