Question:

# How do you subtract a fraction problem give example?

## To subtract 2 fractions with like denominators, just subtract the numerators (example: 3/10-2/10=1/10). If you have different MORE

+
Subtraction

Multiplication
×
Division
÷
Elementary arithmetic is the simplified portion of arithmetic which includes the operations of addition, subtraction, multiplication, and division. Elementary arithmetic starts with the natural numbers and the written symbols (digits) which represent them. The process for combining a pair of these numbers with the four basic operations traditionally relies on memorized results for small values of numbers, including the contents of a multiplication table to assist with multiplication and division. Elementary arithmetic also includes fractions and negative numbers, which can be represented on a number line. The abacus is an early mechanical device for performing elementary arithmetic, which is still used in many parts of Asia. Modern calculating tools which perform elementary arithmetic operations include cash registers, electronic calculators, and computers. Digits are the entire set of symbols used to represent numbers. In a particular numeral system, a single digit represents a different amount than any other digit, although the symbols in the same numeral system might vary between cultures. In modern usage, the Arabic numerals are the most common set of symbols, and the most frequently used form of these digits is the Western style. Each single digit matches the following amounts:
0, zero. Used in the absence of objects to be counted. For example, a different way of saying "there are no sticks here", is to say "the number of sticks here is 0".
1, one. Applied to a single item. For example, here is one stick: I
2, two. Applied to a pair of items. Here are two sticks: I I
3, three. Applied to three items. Here are three sticks: I I I
4, four. Applied to four items. Here are four sticks: I I I  I
5, five. Applied to five items. Here are five sticks: I I I  I I
6, six. Applied to six items. Here are six sticks: I I I  I I I
7, seven. Applied to seven items. Here are seven sticks: I I I  I I I  I
8, eight. Applied to eight items. Here are eight sticks: I I I  I I I  I I
the successor of zero is one,
the successor of one is two,
the successor of two is three,
the successor of ten is eleven.
Thus the product of five and three is fifteen.
+
Subtraction

Multiplication
×
Division
÷
An Egyptian fraction is the sum of distinct unit fractions, such as $\tfrac{1}{2}+\tfrac{1}{3}+\tfrac{1}{16}$. That is, each fraction in the expression has a numerator equal to 1 and a denominator that is a positive integer, and all the denominators differ from each other. The value of an expression of this type is a positive rational number a/b; for instance the Egyptian fraction above sums to 43/48. Every positive rational number can be represented by an Egyptian fraction. Sums of this type, and similar sums also including 2/3 and 3/4 as summands, were used as a serious notation for rational numbers by the ancient Egyptians, and continued to be used by other civilizations into medieval times. In modern mathematical notation, Egyptian fractions have been superseded by vulgar fractions and decimal notation. However, Egyptian fractions continue to be an object of study in modern number theory and recreational mathematics, as well as in modern historical studies of ancient mathematics. Egyptian fraction notation was developed in the Middle Kingdom of Egypt, altering the Old Kingdom's Eye of Horus numeration system. Five early texts in which Egyptian fractions appear were the Egyptian Mathematical Leather Roll, the Moscow Mathematical Papyrus, the Reisner Papyrus, the Kahun Papyrus and the Akhmim Wooden Tablet. A later text, the Rhind Mathematical Papyrus, introduced improved ways of writing Egyptian fractions. The Rhind papyrus was written by Ahmes and dates from the Second Intermediate Period; it includes a ntable of Egyptian fraction expansions for rational numbers 2/, as well as 84 word problems. Solutions to each problem were written out in scribal shorthand, with the final answers of all 84 problems being expressed in Egyptian fraction notation. 2/n tables similar to the one on the Rhind papyrus also appear on some of the other texts. However, as the Kahun Papyrus shows, vulgar fractions were also used by scribes within their calculations. To write the unit fractions used in their Egyptian fraction notation, in hieroglyph script, the Egyptians placed the hieroglyph (er, "[one] among" or possibly re, mouth) above a number to represent the reciprocal of that number. Similarly in hieratic script they drew a line over the letter representing the number. For example: The Egyptians had special symbols for 1/2, 2/3, and 3/4 that were used to reduce the size of numbers greater than 1/2 when such numbers were converted to an Egyptian fraction series. The remaining number after subtracting one of these special fractions was written using as a sum of distinct unit fractions according to the usual Egyptian fraction notation. The Egyptians also used an alternative notation modified from the Old Kingdom and based on the parts of the Eye of Horus to denote a special set of fractions of the form 1/2k (for k = 1, 2, ..., 6) and sums of these numbers, which are necessarily dyadic rational numbers. These "Horus-Eye fractions" were used in the Middle Kingdom in conjunction with the later notation for Egyptian fractions to subdivide a hekat, the primary ancient Egyptian volume measure for grain, bread, and other small quantities of volume, as described in the Akhmim Wooden Tablet. If any remainder was left after expressing a quantity in Eye of Horus fractions of a hekat, the remainder was written using the usual Egyptian fraction notation as multiples of a ro, a unit equal to 1/320 of a hekat. Modern historians of mathematics have studied the Rhind papyrus and other ancient sources in an attempt to discover the methods the Egyptians used in calculating with Egyptian fractions. In particular, study in this area has concentrated on understanding the tables of expansions for numbers of the form 2/n in the Rhind papyrus. Although these expansions can generally be described as algebraic identities, the methods used by the Egyptians may not correspond directly to these identities. Additionally, the expansions in the table do not match any single identity; rather, different identities match the expansions for prime and for composite denominators, and more than one identity fits the numbers of each type: Egyptian fraction notation continued to be used in Greek times and into the Middle Ages (Struik 1967), despite complaints as early as Ptolemy's Almagest about the clumsiness of the notation compared to alternatives such as the Babylonian base-60 notation. An important text of medieval mathematics, the Liber Abaci (1202) of Leonardo of Pisa (more commonly known as Fibonacci), provides some insight into the uses of Egyptian fractions in the Middle Ages, and introduces topics that continue to be important in modern mathematical study of these series. The primary subject of the Liber Abaci is calculations involving decimal and vulgar fraction notation, which eventually replaced Egyptian fractions. Fibonacci himself used a complex notation for fractions involving a combination of a mixed radix notation with sums of fractions. Many of the calculations throughout Fibonacci's book involve numbers represented as Egyptian fractions, and one section of this book (Sigler 2002, chapter II.7) provides a list of methods for conversion of vulgar fractions to Egyptian fractions. If the number is not already a unit fraction, the first method in this list is to attempt to split the numerator into a sum of divisors of the denominator; this is possible whenever the denominator is a practical number, and Liber Abaci includes tables of expansions of this type for the practical numbers 6, 8, 12, 20, 24, 60, and 100. The next several methods involve algebraic identities such as $\tfrac{a}{ab-1}=\tfrac{1}{b}+\tfrac{1}{b(ab-1)}.$ For instance, Fibonacci represents the fraction $\tfrac{8}{11}$ by splitting the numerator into a sum of two numbers, each of which divides one plus the denominator: $\tfrac{8}{11}=\tfrac{6}{11}+\tfrac{2}{11}.$ Fibonacci applies the algebraic identity above to each these two parts, producing the expansion $\tfrac{8}{11}=\tfrac{1}{2}+\tfrac{1}{22}+\tfrac{1}{6}+\tfrac{1}{66}.$ Fibonacci describes similar methods for denominators that are two or three less than a number with many factors. In the rare case that these other methods all fail, Fibonacci suggests a greedy algorithm for computing Egyptian fractions, in which one repeatedly chooses the unit fraction with the smallest denominator that is no larger than the remaining fraction to be expanded: that is, in more modern notation, we replace a fraction x/y by the expansion where $\lceil \ldots \rceil$ represents the ceiling function. Fibonacci suggests switching to another method after the first such expansion, but he also gives examples in which this greedy expansion was iterated until a complete Egyptian fraction expansion was constructed: $\tfrac{4}{13}=\tfrac{1}{4}+\tfrac{1}{18}+\tfrac{1}{468}$ and $\tfrac{17}{29}=\tfrac{1}{2}+\tfrac{1}{12}+\tfrac{1}{348}.$ As later mathematicians showed, each greedy expansion reduces the numerator of the remaining fraction to be expanded, so this method always terminates with a finite expansion. However, compared to ancient Egyptian expansions or to more modern methods, this method may produce expansions that are quite long, with large denominators, and Fibonacci himself noted the awkwardness of the expansions produced by this method. For instance, the greedy method expands while other methods lead to the much better expansion Sylvester's sequence 2, 3, 7, 43, 1807, ... can be viewed as generated by an infinite greedy expansion of this type for the number one, where at each step we choose the denominator $\lfloor y/x\rfloor+1$ instead of $\lceil y/x\rceil$, and sometimes Fibonacci's greedy algorithm is attributed to Sylvester. After his description of the greedy algorithm, Fibonacci suggests yet another method, expanding a fraction $a/b$ by searching for a number c having many divisors, with $b/2 < c < b$, replacing $a/b$ by $ac/bc$, and expanding $ac$ as a sum of divisors of $bc$, similar to the method proposed by Hultsch and Bruins to explain some of the expansions in the Rhind papyrus. Modern number theorists have studied many different problems related to Egyptian fractions, including problems of bounding the length or maximum denominator in Egyptian fraction representations, finding expansions of certain special forms or in which the denominators are all of some special type, the termination of various methods for Egyptian fraction expansion, and showing that expansions exist for any sufficiently dense set of sufficiently smooth numbers. Some notable problems remain unsolved with regard to Egyptian fractions, despite considerable effort by mathematicians. Guy (2004) describes these problems in more detail and lists numerous additional open problems.
In mathematics, especially in elementary arithmetic, division (÷) is an arithmetic operation. Specifically, if b times c equals a, written: where b is not zero, then a divided by b equals c, written: For instance, since In the expression a ÷ b = c, a is called the dividend or numerator, b the divisor or denominator and the result c is called the quotient. Conceptually, division describes two distinct but related settings. Partitioning involves taking a set of size a and forming b groups that are equal in size. The size of each group formed, c, is the quotient of a and b. Quotative division involves taking a set of size a and forming groups of size c. The number of groups of this size that can be formed, b, is the quotient of a and c. Teaching division usually leads to the concept of fractions being introduced to students. Unlike addition, subtraction, and multiplication, the set of all integers is not closed under division. Dividing two integers may result in a remainder. To complete the division of the remainder, the number system is extended to include fractions or rational numbers as they are more generally called. Division is often shown in algebra and science by placing the dividend over the divisor with a horizontal line, also called a vinculum or fraction bar, between them. For example, a divided by b is written This can be read out loud as "a divided by b", "a by b" or "a over b". A way to express division all on one line is to write the dividend (or numerator), then a slash, then the divisor (or denominator), like this: This is the usual way to specify division in most computer programming languages since it can easily be typed as a simple sequence of ASCII characters. A typographical variation halfway between these two forms uses a solidus (fraction slash) but elevates the dividend, and lowers the divisor: Any of these forms can be used to display a fraction. A fraction is a division expression where both dividend and divisor are integers (although typically called the numerator and denominator), and there is no implication that the division must be evaluated further. A second way to show division is to use the obelus (or division sign), common in arithmetic, in this manner: This form is infrequent except in elementary arithmetic. ISO 80000-2-9.6 states it should not be used. The obelus is also used alone to represent the division operation itself, as for instance as a label on a key of a calculator. In some non-English-speaking cultures, "a divided by b" is written a : b. This notation was introduced in 1631 by William Oughtred in his Clavis Mathematicae and later popularized by Gottfried Wilhelm Leibniz. However, in English usage the colon is restricted to expressing the related concept of ratios (then "a is to b"). In elementary mathematics the notation $b)~a$ or $b)\overline{~a~}$ is used to denote a divided by b. This notation was first introduced by Michael Stifel in Arithmetica integra, published in 1544. Division is often introduced through the notion of "sharing out" a set of objects, for example a pile of sweets, into a number of equal portions. Distributing the objects several at a time in each round of sharing to each portion leads to the idea of "chunking", i.e., division by repeated subtraction. More systematic and more efficient (but also more formalised and more rule-based, and more removed from an overall holistic picture of what division is achieving), a person who knows the multiplication tables can divide two integers using pencil and paper using the method of short division, if the divisor is simple. Long division is used for larger integer divisors. If the dividend has a fractional part (expressed as a decimal fraction), one can continue the algorithm past the ones place as far as desired. If the divisor has a fractional part, we can restate the problem by moving the decimal to the right in both numbers until the divisor has no fraction. A person can calculate division with an abacus by repeatedly placing the dividend on the abacus, and then subtracting the divisor the offset of each digit in the result, counting the number of divisions possible at each offset. A person can use logarithm tables to divide two numbers, by subtracting the two numbers' logarithms, then looking up the antilogarithm of the result. A person can calculate division with a slide rule by aligning the divisor on the C scale with the dividend on the D scale. The quotient can be found on the D scale where it is aligned with the left index on the C scale. The user is responsible, however, for mentally keeping track of the decimal point. Modern computers compute division by methods that are faster than long division: see Division algorithm. In modular arithmetic, some numbers have a multiplicative inverse with respect to the modulus. We can calculate division by multiplication in such a case. This approach is useful in computers that do not have a fast division instruction. The division algorithm is a mathematical theorem that precisely expresses the outcome of the usual process of division of integers. In particular, the theorem asserts that integers called the quotient q and remainder r always exist and that they are uniquely determined by the dividend a and divisor d, with d ≠ 0. Formally, the theorem is stated as follows: There exist unique integers q and r such that a = qd + r and 0 ≤ r < | d |, where | d | denotes the absolute value of d. Division of integers is not closed. Apart from division by zero being undefined, the quotient is not an integer unless the dividend is an integer multiple of the divisor. For example 26 cannot be divided by 11 to give an integer. Such a case uses one of five approaches: Dividing integers in a computer program requires special care. Some programming languages, such as C, treat integer division as in case 5 above, so the answer is an integer. Other languages, such as MATLAB and every computer algebra system return a rational number as the answer, as in case 3 above. These languages provide also functions to get the results of the other cases, either directly of from the result of case 3. Names and symbols used for integer division include div, /, \, and %. Definitions vary regarding integer division when the dividend or the divisor is negative: rounding may be toward zero (so called T-division) or toward −∞ (F-division); rarer styles can occur – see Modulo operation for the details. Divisibility rules can sometimes be used to quickly determine whether one integer divides exactly into another. The result of dividing two rational numbers is another rational number when the divisor is not 0. We may define division of two rational numbers p/q and r/s by All four quantities are integers, and only p may be 0. This definition ensures that division is the inverse operation of multiplication. Division of two real numbers results in another real number when the divisor is not 0. It is defined such a/b = c if and only if a = cb and b ≠ 0. Division of any number by zero (where the divisor is zero) is undefined. This is because zero multiplied by any finite number always results in a product of zero. Entry of such an expression into most calculators produces an error message. Dividing two complex numbers results in another complex number when the divisor is not 0, defined thus: All four quantities are real numbers. r and s may not both be 0. Division for complex numbers expressed in polar form is simpler than the definition above: Again all four quantities are real numbers. r may not be 0. One can define the division operation for polynomials in one variable over a field. Then, as in the case of integers, one has a remainder. See Euclidean division of polynomials, and, for hand-written computation, polynomial long division or synthetic division. One can define a division operation for matrices. The usual way to do this is to define , where denotes the inverse of B, but it is far more common to write out explicitly to avoid confusion. Because matrix multiplication is not commutative, one can also define a left division or so-called backslash-division as . For this to be well defined, need not exist, however does need to exist. To avoid confusion, division as defined by is sometimes called right division or slash-division in this context. Note that with left and right division defined this way, is in general not the same as and nor is the same as , but and . To avoid problems when and/or do not exist, division can also be defined as multiplication with the pseudoinverse, i.e., and , where and denote the pseudoinverse of A and B. In abstract algebras such as matrix algebras and quaternion algebras, fractions such as ${a \over b}$ are typically defined as $a \cdot {1 \over b}$ or $a \cdot b^{-1}$ where $b$ is presumed an invertible element (i.e., there exists a multiplicative inverse $b^{-1}$ such that $bb^{-1} = b^{-1}b = 1$ where $1$ is the multiplicative identity). In an integral domain where such elements may not exist, division can still be performed on equations of the form $ab = ac$ or $ba = ca$ by left or right cancellation, respectively. More generally "division" in the sense of "cancellation" can be done in any ring with the aforementioned cancellation properties. If such a ring is finite, then by an application of the pigeonhole principle, every nonzero element of the ring is invertible, so division by any nonzero element is possible in such a ring. To learn about when algebras (in the technical sense) have a division operation, refer to the page on division algebras. In particular Bott periodicity can be used to show that any real normed division algebra must be isomorphic to either the real numbers R, the complex numbers C, the quaternions H, or the octonions O. The derivative of the quotient of two functions is given by the quotient rule: There is no general method to integrate the quotient of two functions.
+
Subtraction

Multiplication
×
Division
÷
A unit fraction is a rational number written as a fraction where the numerator is one and the denominator is a positive integer. A unit fraction is therefore the reciprocal of a positive integer, 1/n. Examples are 1/1, 1/2, 1/3, 1/4 etc. Multiplying any two unit fractions results in a product that is another unit fraction: However, adding, subtracting, or dividing two unit fractions produces a result that is generally not a unit fraction: Unit fractions play an important role in modular arithmetic, as they may be used to reduce modular division to the calculation of greatest common divisors. Specifically, suppose that we wish to perform divisions by a value x, modulo y. In order for division by x to be well defined modulo y, x and y must be relatively prime. Then, by using the extended Euclidean algorithm for greatest common divisors we may find a and b such that from which it follows that or equivalently Thus, to divide by x (modulo y) we need merely instead multiply by a. Any positive rational number can be written as the sum of unit fractions, in multiple ways. For example, The ancient Egyptians used sums of distinct unit fractions in their notation for more general rational numbers, and so such sums are often called Egyptian fractions. There is still interest today in analyzing the methods used by the ancients to choose among the possible representations for a fractional number, and to calculate with such representations. The topic of Egyptian fractions has also seen interest in modern number theory; for instance, the Erdős–Graham conjecture and the Erdős–Straus conjecture concern sums of unit fractions, as does the definition of Ore's harmonic numbers. In geometric group theory, triangle groups are classified into Euclidean, spherical, and hyperbolic cases according to whether an associated sum of unit fractions is equal to one, greater than one, or less than one respectively. Many well-known infinite series have terms that are unit fractions. These include: The Hilbert matrix is the matrix with elements It has the unusual property that all elements in its inverse matrix are integers. Similarly, Richardson (2001) defined a matrix with elements where Fi denotes the ith Fibonacci number. He calls this matrix the Filbert matrix and it has the same property of having an integer inverse. Two fractions are called adjacent if their difference is a unit fraction. In a uniform distribution on a discrete space, all probabilities are equal unit fractions. Due to the principle of indifference, probabilities of this form arise frequently in statistical calculations. Additionally, Zipf's law states that, for many observed phenomena involving the selection of items from an ordered sequence, the probability that the nth item is selected is proportional to the unit fraction 1/n. The energy levels of photons that can be absorbed or emitted by a hydrogen atom are, according to the Rydberg formula, proportional to the differences of two unit fractions. An explanation for this phenomenon is provided by the Bohr model, according to which the energy levels of electron orbitals in a hydrogen atom are inversely proportional to square unit fractions, and the energy of a photon is quantized to the difference between two levels. Arthur Eddington argued that the fine structure constant was a unit fraction, first 1/136 then 1/137. This contention has been falsified, given that current estimates of the fine structure constant are (to 6 significant digits) 1/137.036.
Addition is used to model countless physical processes. Even for the simple case of adding natural numbers, there are many possible interpretations and even more visual representations. Possibly the most fundamental interpretation of addition lies in combining sets: This interpretation is easy to visualize, with little danger of ambiguity. It is also useful in higher mathematics; for the rigorous definition it inspires, see Natural numbers below. However, it is not obvious how one should extend this version of addition to include fractional numbers or negative numbers. One possible fix is to consider collections of objects that can be easily divided, such as pies or, still better, segmented rods. Rather than just combining collections of segments, rods can be joined end-to-end, which illustrates another conception of addition: adding not the rods but the lengths of the rods. A second interpretation of addition comes from extending an initial length by a given length: The sum a + b can be interpreted as a binary operation that combines a and b, in an algebraic sense, or it can be interpreted as the addition of b more units to a. Under the latter interpretation, the parts of a sum a + b play asymmetric roles, and the operation a + b is viewed as applying the unary operation +b to a. Instead of calling both a and b addends, it is more appropriate to call a the augend in this case, since a plays a passive role. The unary view is also useful when discussing subtraction, because each unary addition operation has an inverse unary subtraction operation, and vice versa. Addition is commutative, meaning that one can reverse the terms in a sum left-to-right, and the result is the same as the last one. Symbolically, if a and b are any two numbers, then The fact that addition is commutative is known as the "commutative law of addition". This phrase suggests that there are other commutative laws: for example, there is a commutative law of multiplication. However, many binary operations are not commutative, such as subtraction and division, so it is misleading to speak of an unqualified "commutative law". A somewhat subtler property of addition is associativity, which comes up when one tries to define repeated addition. Should the expression be defined to mean (a + b) + c or a + (b + c)? That addition is associative tells us that the choice of definition is irrelevant. For any three numbers a, b, and c, it is true that For example, (1 + 2) + 3 = 3 + 3 = 6 = 1 + 5 = 1 + (2 + 3). Not all operations are associative, so in expressions with other operations like subtraction, it is important to specify the order of operations. When adding zero to any number, the quantity does not change; zero is the identity element for addition, also known as the additive identity. In symbols, for any a, This law was first identified in Brahmagupta's Brahmasphutasiddhanta in 628 AD, although he wrote it as three separate laws, depending on whether a is negative, positive, or zero itself, and he used words rather than algebraic symbols. Later Indian mathematicians refined the concept; around the year 830, Mahavira wrote, "zero becomes the same as what is added to it", corresponding to the unary statement 0 + a = a. In the 12th century, Bhaskara wrote, "In the addition of cipher, or subtraction of it, the quantity, positive or negative, remains the same", corresponding to the unary statement a + 0 = a. In the context of integers, addition of one also plays a special role: for any integer a, the integer (a + 1) is the least integer greater than a, also known as the successor of a. Because of this succession, the value of some a + b can also be seen as the $b^{th}$ successor of a, making addition iterated succession. To numerically add physical quantities with units, they must first be expressed with common units. For example, if a measure of 5 feet is extended by 2 inches, the sum is 62 inches, since 60 inches is synonymous with 5 feet. On the other hand, it is usually meaningless to try to add 3 meters and 4 square meters, since those units are incomparable; this sort of consideration is fundamental in dimensional analysis. Studies on mathematical development starting around the 1980s have exploited the phenomenon of habituation: infants look longer at situations that are unexpected. A seminal experiment by Karen Wynn in 1992 involving Mickey Mouse dolls manipulated behind a screen demonstrated that five-month-old infants expect 1 + 1 to be 2, and they are comparatively surprised when a physical situation seems to imply that 1 + 1 is either 1 or 3. This finding has since been affirmed by a variety of laboratories using different methodologies. Another 1992 experiment with older toddlers, between 18 to 35 months, exploited their development of motor control by allowing them to retrieve ping-pong balls from a box; the youngest responded well for small numbers, while older subjects were able to compute sums up to 5. Even some nonhuman animals show a limited ability to add, particularly primates. In a 1995 experiment imitating Wynn's 1992 result (but using eggplants instead of dolls), rhesus macaques and cottontop tamarins performed similarly to human infants. More dramatically, after being taught the meanings of the Arabic numerals 0 through 4, one chimpanzee was able to compute the sum of two numerals without further training. Typically, children first master counting. When given a problem that requires that two items and three items be combined, young children model the situation with physical objects, often fingers or a drawing, and then count the total. As they gain experience, they learn or discover the strategy of "counting-on": asked to find two plus three, children count three past two, saying "three, four, five" (usually ticking off fingers), and arriving at five. This strategy seems almost universal; children can easily pick it up from peers or teachers. Most discover it independently. With additional experience, children learn to add more quickly by exploiting the commutativity of addition by counting up from the larger number, in this case starting with three and counting "four, five." Eventually children begin to recall certain addition facts ("number bonds"), either through experience or rote memorization. Once some facts are committed to memory, children begin to derive unknown facts from known ones. For example, a child asked to add six and seven may know that 6+6=12 and then reason that 6+7 is one more, or 13. Such derived facts can be found very quickly and most elementary school student eventually rely on a mixture of memorized and derived facts to add fluently. The prerequisite to addition in the decimal system is the fluent recall or derivation of the 100 single-digit "addition facts". One could memorize all the facts by rote, but pattern-based strategies are more enlightening and, for most people, more efficient: As students grow older, they commit more facts to memory, and learn to derive other facts rapidly and fluently. Many students never commit all the facts to memory, but can still find any basic fact quickly. The standard algorithm for adding multidigit numbers is to align the addends vertically and add the columns, starting from the ones column on the right. If a column exceeds ten, the extra digit is "carried" into the next column. An alternate strategy starts adding from the most significant digit on the left; this route makes carrying a little clumsier, but it is faster at getting a rough estimate of the sum. There are many other alternative methods. Analog computers work directly with physical quantities, so their addition mechanisms depend on the form of the addends. A mechanical adder might represent two addends as the positions of sliding blocks, in which case they can be added with an averaging lever. If the addends are the rotation speeds of two shafts, they can be added with a differential. A hydraulic adder can add the pressures in two chambers by exploiting Newton's second law to balance forces on an assembly of pistons. The most common situation for a general-purpose analog computer is to add two voltages (referenced to ground); this can be accomplished roughly with a resistor network, but a better design exploits an operational amplifier. Addition is also fundamental to the operation of digital computers, where the efficiency of addition, in particular the carry mechanism, is an important limitation to overall performance. Blaise Pascal invented the mechanical calculator in 1642, it was the first operational adding machine. It made use of an ingenious gravity-assisted carry mechanism. It was the only operational mechanical calculator in the 17th century and the earliest automatic, digital computers. Pascal's calculator was limited by its carry mechanism which forced its wheels to only turn one way, so it could add but, to subtract, the operator had to use of the method of complements which required as many steps as an addition. Pascal was followed by Giovanni Poleni who built the second functional mechanical calculator in 1709, a calculating clock, which was made of wood and which could, once setup, multiply two numbers automatically. Adders execute integer addition in electronic digital computers, usually using binary arithmetic. The simplest architecture is the ripple carry adder, which follows the standard multi-digit algorithm. One slight improvement is the carry skip design, again following human intuition; one does not perform all the carries in computing 999 + 1, but one bypasses the group of 9s and skips to the answer. Since they compute digits one at a time, the above methods are too slow for most modern purposes. In modern digital computers, integer addition is typically the fastest arithmetic instruction, yet it has the largest impact on performance, since it underlies all the floating-point operations as well as such basic tasks as address generation during memory access and fetching instructions during branching. To increase speed, modern designs calculate digits in parallel; these schemes go by such names as carry select, carry lookahead, and the Ling pseudocarry. Almost all modern implementations are, in fact, hybrids of these last three designs. Unlike addition on paper, addition on a computer often changes the addends. On the ancient abacus and adding board, both addends are destroyed, leaving only the sum. The influence of the abacus on mathematical thinking was strong enough that early Latin texts often claimed that in the process of adding "a number to a number", both numbers vanish. In modern times, the ADD instruction of a microprocessor replaces the augend with the sum but preserves the addend. In a high-level programming language, evaluating a + b does not change either a or b; if the goal is to replace a with the sum this must be explicitly requested, typically with the statement a = a + b. Some languages such as C or C++ allow this to be abbreviated as a += b. To prove the usual properties of addition, one must first define addition for the context in question. Addition is first defined on the natural numbers. In set theory, addition is then extended to progressively larger sets that include the natural numbers: the integers, the rational numbers, and the real numbers. (In mathematics education, positive fractions are added before negative numbers are even considered; this is also the historical route) There are two popular ways to define the sum of two natural numbers a and b. If one defines natural numbers to be the cardinalities of finite sets, (the cardinality of a set is the number of elements in the set), then it is appropriate to define their sum as follows: Here, A U B is the union of A and B. An alternate version of this definition allows A and B to possibly overlap and then takes their disjoint union, a mechanism that allows common elements to be separated out and therefore counted twice. The other popular definition is recursive: Again, there are minor variations upon this definition in the literature. Taken literally, the above definition is an application of the Recursion Theorem on the poset N2. On the other hand, some sources prefer to use a restricted Recursion Theorem that applies only to the set of natural numbers. One then considers a to be temporarily "fixed", applies recursion on b to define a function "a + ", and pastes these unary operations for all a together to form the full binary operation. This recursive formulation of addition was developed by Dedekind as early as 1854, and he would expand upon it in the following decades. He proved the associative and commutative properties, among others, through mathematical induction; for examples of such inductive proofs, see Addition of natural numbers. The simplest conception of an integer is that it consists of an absolute value (which is a natural number) and a sign (generally either positive or negative). The integer zero is a special third case, being neither positive nor negative. The corresponding definition of addition must proceed by cases: Although this definition can be useful for concrete problems, it is far too complicated to produce elegant general proofs; there are too many cases to consider. A much more convenient conception of the integers is the Grothendieck group construction. The essential observation is that every integer can be expressed (not uniquely) as the difference of two natural numbers, so we may as well define an integer as the difference of two natural numbers. Addition is then defined to be compatible with subtraction: Addition of rational numbers can be computed using the least common denominator, but a conceptually simpler definition involves only integer addition and multiplication: The commutativity and associativity of rational addition is an easy consequence of the laws of integer arithmetic. For a more rigorous and general discussion, see field of fractions. A common construction of the set of real numbers is the Dedekind completion of the set of rational numbers. A real number is defined to be a Dedekind cut of rationals: a non-empty set of rationals that is closed downward and has no greatest element. The sum of real numbers a and b is defined element by element: This definition was first published, in a slightly modified form, by Richard Dedekind in 1872. The commutativity and associativity of real addition are immediate; defining the real number 0 to be the set of negative rationals, it is easily seen to be the additive identity. Probably the trickiest part of this construction pertaining to addition is the definition of additive inverses. Unfortunately, dealing with multiplication of Dedekind cuts is a case-by-case nightmare similar to the addition of signed integers. Another approach is the metric completion of the rational numbers. A real number is essentially defined to be the a limit of a Cauchy sequence of rationals, lim an. Addition is defined term by term: This definition was first published by Georg Cantor, also in 1872, although his formalism was slightly different. One must prove that this operation is well-defined, dealing with co-Cauchy sequences. Once that task is done, all the properties of real addition follow immediately from the properties of rational numbers. Furthermore, the other arithmetic operations, including multiplication, have straightforward, analogous definitions. There are many binary operations that can be viewed as generalizations of the addition operation on the real numbers. The field of abstract algebra is centrally concerned with such generalized operations, and they also appear in set theory and category theory. In linear algebra, a vector space is an algebraic structure that allows for adding any two vectors and for scaling vectors. A familiar vector space is the set of all ordered pairs of real numbers; the ordered pair (a,b) is interpreted as a vector from the origin in the Euclidean plane to the point (a,b) in the plane. The sum of two vectors is obtained by adding their individual coordinates: This addition operation is central to classical mechanics, in which vectors are interpreted as forces. In modular arithmetic, the set of integers modulo 12 has twelve elements; it inherits an addition operation from the integers that is central to musical set theory. The set of integers modulo 2 has just two elements; the addition operation it inherits is known in Boolean logic as the "exclusive or" function. In geometry, the sum of two angle measures is often taken to be their sum as real numbers modulo 2π. This amounts to an addition operation on the circle, which in turn generalizes to addition operations on many-dimensional tori. The general theory of abstract algebra allows an "addition" operation to be any associative and commutative operation on a set. Basic algebraic structures with such an addition operation include commutative monoids and abelian groups. A far-reaching generalization of addition of natural numbers is the addition of ordinal numbers and cardinal numbers in set theory. These give two different generalizations of addition of natural numbers to the transfinite. Unlike most addition operations, addition of ordinal numbers is not commutative. Addition of cardinal numbers, however, is a commutative operation closely related to the disjoint union operation. In category theory, disjoint union is seen as a particular case of the coproduct operation, and general coproducts are perhaps the most abstract of all the generalizations of addition. Some coproducts, such as Direct sum and Wedge sum, are named to evoke their connection with addition. Subtraction can be thought of as a kind of addition—that is, the addition of an additive inverse. Subtraction is itself a sort of inverse to addition, in that adding x and subtracting x are inverse functions. Given a set with an addition operation, one cannot always define a corresponding subtraction operation on that set; the set of natural numbers is a simple example. On the other hand, a subtraction operation uniquely determines an addition operation, an additive inverse operation, and an additive identity; for this reason, an additive group can be described as a set that is closed under subtraction. Multiplication can be thought of as repeated addition. If a single term x appears in a sum n times, then the sum is the product of n and x. If n is not a natural number, the product may still make sense; for example, multiplication by −1 yields the additive inverse of a number. In the real and complex numbers, addition and multiplication can be interchanged by the exponential function: This identity allows multiplication to be carried out by consulting a table of logarithms and computing addition by hand; it also enables multiplication on a slide rule. The formula is still a good first-order approximation in the broad context of Lie groups, where it relates multiplication of infinitesimal group elements with addition of vectors in the associated Lie algebra. There are even more generalizations of multiplication than addition. In general, multiplication operations always distribute over addition; this requirement is formalized in the definition of a ring. In some contexts, such as the integers, distributivity over addition and the existence of a multiplicative identity is enough to uniquely determine the multiplication operation. The distributive property also provides information about addition; by expanding the product (1 + 1)(a + b) in both ways, one concludes that addition is forced to be commutative. For this reason, ring addition is commutative in general. Division is an arithmetic operation remotely related to addition. Since a/b = a(b−1), division is right distributive over addition: (a + b) / c = a / c + b / c. However, division is not left distributive over addition; 1/ (2 + 2) is not the same as 1/2 + 1/2. The maximum operation "max (a, b)" is a binary operation similar to addition. In fact, if two nonnegative numbers a and b are of different orders of magnitude, then their sum is approximately equal to their maximum. This approximation is extremely useful in the applications of mathematics, for example in truncating Taylor series. However, it presents a perpetual difficulty in numerical analysis, essentially since "max" is not invertible. If b is much greater than a, then a straightforward calculation of (a + b) − b can accumulate an unacceptable round-off error, perhaps even returning zero. See also Loss of significance. The approximation becomes exact in a kind of infinite limit; if either a or b is an infinite cardinal number, their cardinal sum is exactly equal to the greater of the two. Accordingly, there is no subtraction operation for infinite cardinals. Maximization is commutative and associative, like addition. Furthermore, since addition preserves the ordering of real numbers, addition distributes over "max" in the same way that multiplication distributes over addition: For these reasons, in tropical geometry one replaces multiplication with addition and addition with maximization. In this context, addition is called "tropical multiplication", maximization is called "tropical addition", and the tropical "additive identity" is negative infinity. Some authors prefer to replace addition with minimization; then the additive identity is positive infinity. Tying these observations together, tropical addition is approximately related to regular addition through the logarithm: which becomes more accurate as the base of the logarithm increases. The approximation can be made exact by extracting a constant h, named by analogy with Planck's constant from quantum mechanics, and taking the "classical limit" as h tends to zero: In this sense, the maximum operation is a dequantized version of addition. Incrementation, also known as the successor operation, is the addition of 1 to a number. Summation describes the addition of arbitrarily many numbers, usually more than just two. It includes the idea of the sum of a single number, which is itself, and the empty sum, which is zero. An infinite summation is a delicate procedure known as a series. Counting a finite set is equivalent to summing 1 over the set. Integration is a kind of "summation" over a continuum, or more precisely and generally, over a differentiable manifold. Integration over a zero-dimensional manifold reduces to summation. Linear combinations combine multiplication and summation; they are sums in which each term has a multiplier, usually a real or complex number. Linear combinations are especially useful in contexts where straightforward addition would violate some normalization rule, such as mixing of strategies in game theory or superposition of states in quantum mechanics. Convolution is used to add two independent random variables defined by distribution functions. Its usual definition combines integration, subtraction, and multiplication. In general, convolution is useful as a kind of domain-side addition; by contrast, vector addition is a kind of range-side addition.
+
Subtraction

Multiplication
×
Division
÷

Mathematical analysis is a branch of mathematics that includes the theories of differentiation, integration, measure, limits, infinite series, and analytic functions. These theories are usually studied in the context of real and complex numbers and functions. Analysis evolved from calculus, which involves the elementary concepts and techniques of analysis. Analysis may be distinguished from geometry. However, it can be applied to any space of mathematical objects that has a definition of nearness (a topological space) or specific distances between objects (a metric space).

Early results in analysis were implicitly present in the early days of ancient Greek mathematics. For instance, an infinite geometric sum is implicit in Zeno's paradox of the dichotomy. Later, Greek mathematicians such as Eudoxus and Archimedes made more explicit, but informal, use of the concepts of limits and convergence when they used the method of exhaustion to compute the area and volume of regions and solids. In India, the 12th century mathematician Bhāskara II gave examples of the derivative and used what is now known as Rolle's theorem.

Division

Elementary arithmetic is the simplified portion of arithmetic which includes the operations of addition, subtraction, multiplication, and division.

Elementary arithmetic starts with the natural numbers and the written symbols (digits) which represent them. The process for combining a pair of these numbers with the four basic operations traditionally relies on memorized results for small values of numbers, including the contents of a multiplication table to assist with multiplication and division.

Fraction Numbers

Quote notation is a number system for representing rational numbers which was designed to be attractive for use in computer architecture. In a typical computer architecture, the representation and manipulation of rational numbers is a complex topic. In Quote notation, arithmetic operations take particularly simple, consistent forms, and can produce exact answers with no roundoff error.

Quote notation’s arithmetic algorithms work with a typical right-to-left direction, in which the addition, subtraction, and multiplication algorithms have the same complexity for natural numbers, and division is easier than a typical division algorithm.

Mathematics Arithmetic

In mathematics, a continued fraction is an expression obtained through an iterative process of representing a number as the sum of its integer part and the reciprocal of another number, then writing this other number as the sum of its integer part and another reciprocal, and so on. In a finite continued fraction (or terminated continued fraction), the iteration/recursion is terminated after finitely many steps by using an integer in lieu of another continued fraction. In contrast, an infinite continued fraction is an infinite expression. In either case, all integers in the sequence, other than the first, must be positive. The integers ai are called the coefficients or terms of the continued fraction.

Continued fractions have a number of remarkable properties related to the Euclidean algorithm for integers or real numbers. Every rational number p/q has two closely related expressions as a finite continued fraction, whose coefficients ai can be determined by applying the Euclidean algorithm to (p, q). The numerical value of an infinite continued fraction will be irrational; it is defined from its infinite sequence of integers as the limit of a sequence of values for finite continued fractions. Each finite continued fraction of the sequence is obtained by using a finite prefix of the infinite continued fraction's defining sequence of integers. Moreover, every irrational number α is the value of a unique infinite continued fraction, whose coefficients can be found using the non-terminating version of the Euclidean algorithm applied to the incommensurable values α and 1. This way of expressing real numbers (rational and irrational) is called their continued fraction representation.

28