Real Numbers Essay Example
Real Numbers Essay Example

Real Numbers Essay Example

Available Only on StudyHippo
Topics:
View Entire Sample
Text preview

In mathematics, a real number is a value that represents a quantity along a continuum, such as 5 (an integer), 3/4 (a rational number that is not an integer), 8. 6 (a rational number expressed in decimal representation), and π (3. 1415926535..., an irrational number). As a subset of the real numbers, the integers, such as 5, express discrete rather than continuous quantities. Complex numbers include real numbers as a special case. Real numbers can be divided into rational numbers, such as 42 and ? 23/129, and irrational numbers, such as pi and the square root of two. A real number can be given by an infinite decimal representation, such as 2. 4871773339... , where the digits continue indefinitely. The real numbers are sometimes thought of as points on an infinitely long line called the number line or real line.

History

...

Vulgar fractions had been used by the Egyptians around 1000 BC; the Vedic "Sulba Sutras" ("The rules of chords") in, ca. 600 BC, include what may be the first 'use' of irrational numbers. The concept of irrationality was implicitly accepted by early Indian mathematicians since Manava (c. 50–690 BC), who were aware that the square roots of certain numbers such as 2 and 61 could not be exactly determined.Around 500 BC, the Greek mathematicians led by Pythagoras realized the need for irrational numbers, in particular the irrationality of the square root of 2. The Middle Ages saw the acceptance of zero, negative, integral and fractional numbers, first by Indian and Chinese mathematicians, and then by Arabic mathematicians, who were also the first to treat irrational numbers as algebraic objects, which was made possible by the

View entire sample
Join StudyHippo to see entire essay

development of algebra.

Arabic mathematicians merged the concepts of "number" and "magnitude" into a more general idea of real numbers. The Egyptian mathematician Abu Kamil Shuja ibn Aslam (c. 850–930) was the first to accept irrational numbers as solutions to quadratic equations or as coefficients in an equation, often in the form of square roots, cube roots and fourth roots. In the 16th century, Simon Stevin created the basis for modern decimal notation, and insisted that here is no difference between rational and irrational numbers in this regard.In the 17th century, Descartes introduced the term "real" to describe roots of a polynomial, distinguishing them from "imaginary" ones.

In the 18th and 19th centuries there was much work on irrational and transcendental numbers. Johann Heinrich Lambert (1761) gave the first flawed proof that ? cannot be rational; Adrien-Marie Legendre (1794) completed the proof, and showed that ? s not the square root of a rational number. Paolo Ruffini (1799) and Niels Henrik Abel (1842) both constructed proofs of Abel–Ruffini theorem: that the general quintic or higher equations cannot be solved by a general formula involving only arithmetical operations and roots. Evariste Galois (1832) developed techniques for determining whether a given equation could be solved by radicals, which gave rise to the field of Galois theory.Joseph Liouville (1840) showed that neither e nor e2 can be a root of an integer quadratic equation, and then established existence of transcendental numbers, the proof being subsequently displaced by Georg Cantor (1873).

Charles Hermite (1873) first proved that e is transcendental, and Ferdinand von Lindemann (1882), showed that ? is transcendental. Lindemann's proof was much simplified by Weierstrass (1885), still further by David

Hilbert (1893), and has finally been made elementary by Adolf Hurwitz and Paul Gordan.The development of calculus in the 18th century used the entire set of real numbers without having defined them cleanly. The first rigorous definition was given by Georg Cantor in 1871.

In 1874 he showed that the set of all real numbers is uncountably infinite but the set of all algebraic numbers is countably infinite. Contrary to widely held beliefs, his first method was not his famous diagonal argument, which he published in 1891. See Cantor's first uncountability proof.

Get an explanation on any task
Get unstuck with the help of our AI assistant in seconds
New