`and Place
`
`Amelia A Williamson
`
`In the United States, decimals are notated
`with periods (for example, 1.23}, whereas
`most of Europe uses decimal commas
`{1.23). That is iust one instance in which
`style rules in the United States clash with
`those in Europe. How did this style dif—
`ference arise? What issues does it cause
`for editors? How do editors deal with the
`
`difference? Knowing the answers to those
`questions can help editors to avoid mis—
`understandings while working with both
`styles of decimal separators.
`
`History of the Decimal Point
`The history of decimal notation and the
`style difference dates back to the 1500s.
`when decimal Fractions were first used.
`
`Mathematics historian Florian Cajori
`explained the history in his 1928 hook A
`History of Mathematical Notations. 1
`Simon Stevin, a Belgian mathemati—
`cian in the 16th century. is credited with
`establishing the use of decimal fractions.
`He explained their use in his pamphlet
`La Thiende, published in 1585. Stevin
`understood the nature and importance of
`decimal fractions, but his notation was
`awkward. Stevin notated decimal fractions
`
`by putting superscript circled numbers after
`each digit
`in the number. A superscript
`circled 0 was put after the units-place digit,
`a superscript circled 1 after the tenths—
`place digit, a superscript circled 2 after the
`hundredthsvplace digit, and so on.
`Franciscus Vieta, a French mathemati:
`cian in the late 1500s. came close to the
`modem notation of decimals when he used
`
`a vertical stroke to separate the integer part
`of the number from the fractional part.
`
`AMELIA A WILLIAMSON. a graduate student
`in science and technology jimmalism at Texas
`ASM University, prepared this article while
`a Science Editor intern.
`
`to Barth-
`refer
`Historians sometimes
`olomaeus I’itiscus—a German mathemati-
`
`cian, astronomer, and theologian—as the
`first to use the decimal point. Pitiscus did
`use decimal fractions in his writing, and
`there were dots within the numbers in his
`
`text. Some historians argue, however, that
`the dots were not used as decimal points
`and were merely punctuation points written
`before and after numbers in running text, as
`was common in medieval manuscripts.
`Pitiscus used three other notations for
`
`decimal fractions: he put a zero in front of
`the decimal numbers {for example, 0123
`for our 0.123}. he used a vertical stroke as
`a decimal separator (l | 234 for our 1.234),
`and he used the common fraction form
`
`{1 2343'1000 for our 1.234}. The dots in his
`text were used for other purposes.
`john Napier, a Scottish mathematician,
`appears to have been the first to intention-
`ally use the period as a decimal separator
`in his 161? book Rahdologia. In one part of
`the book. he explicitly says that whatever
`is written after the period is a fraction.
`Later in the hook, however, he uses the
`comma as a decimal separator and shows
`a division problem that illustrates his use
`of the comma as a decimal. “Thus. Napier
`vacillated between the period and the
`comma." Cajori stated; “mathematicians
`have been vacillating in this matter ever
`since.”
`
`Other Decimal Notations
`
`During the 17th century. many other notav
`tions were used besides the period and the
`comma. Some mathematicians put
`the
`decimal part in superscript and underlined
`it, some used an “L" shape as a separator,
`and others used the colon or an inverted
`
`and raised comma as a separator.
`By the 18th century, the notations for
`decimal fractions started to converge. “The
`chaos in notations for decimal fractions
`
`gradually gave way to a semblance of order.
`The situation reduced itself to trials of
`
`strength between the comma and the dot
`as [decimal separators]," Cajori stated.
`In the early 17005, Gottfried Wilhelm
`Leibniz, a German polymath, proposed
`the dot as the symbol for multiplication.
`Therefore. most of Europe favored the
`comma as a decimal separator. In England
`at the time, however. the preferred symbol
`for multiplication was an "X", so the dot
`was used more frequently as a decimal
`separator there than in the rest of Europe.
`During the 19th century in England,
`the dot continued to be used as a decimal
`
`separator and also came to he used as the
`symbol for multiplication. That did not
`cause confusion, however, because the dots
`were placed at different heights. The dot
`used to signify multiplication was placed
`on the baseline (where the United States
`currently places decimals}, and the dot
`used as a decimal separator was placed mid-
`way up (where the United States currently
`places the dot for multiplication).
`In the United States, although the
`comma was used occasionally as a decimal
`separator, the baseline dot has always been
`preferred. Historians are not sure why the
`use of the dot raised halfway as a decimal
`separator never gained general adoption in
`the United States.
`For a while,
`the baseline dot had two
`meanings in the United States: a deci'
`mal separator and a multiplication sign.
`Around 1880. however, the need arose for
`a distinction, and the decimal notation for
`the United States was clearly established
`as a dot on the baseline (for example,
`1.234). The dot used as a multiplication
`sign was raised to the central position, as
`it is today.
`That decision has remained the US stan'
`dard ever since. The United States writes
`
`the decimal on the baseline and England in
`the raised position, and the United States
`writes the multiplication dot in the raised
`position and England on the baseline.
`In the early 1900s,
`the American
`
`42 O Science Editor O March — April 2008 O Vol 31 O No 2
`
`WATSON LABORATORIES V. UNITED THERAPEUTICS, |PR2017-O1622
`
`UNITED THERAPEUTiCS, EX. 2116
`
`Page 1 of 2
`
`
`
`Period or Comma continued
`
`Committee of Mathematicians justified
`the style decision. The committee stated
`that because the letter X is used often in
`
`1.234 to mean one thousand, two hundred
`thirty—four instead of the decimal that was
`intended.
`
`is preferable to use the
`it
`mathematics.
`raised dot for multiplication. The commit-
`tee also stated, “[nasmuch as the period
`will continue to be used in this country
`as a decimal point,
`it
`is likely to cause
`confusion, to elementary pupils at least, to
`attempt to use it as a symbol for multiplica—
`tion," according to Cajori.
`
`Current Style and Issues Editors
`Face
`
`Currently, in European countries except for
`the United Kingdom, the comma is used
`as the decimal separator. In the United
`Kingdom. the raised dot is used. and in
`the United States, the baseline dot is used.
`Australia and most Asian countries use the
`dot, South America uses the comma, and
`some parts of Africa use the dot and other
`parts, the comma
`With regard to use of decimals in sci-
`entific writing. one would think that such
`style differences could cause some confu—
`sion. Peggy Robinson, chair of the commit—
`tee that prepared the most recent edition
`of CSE’s Scientific Style and Format. says
`that the style difference is most likely to
`cause confusion among readers who are
`not Familiar with the difference in con—
`
`vention. An American reading a French
`article,
`for example, might misinterpret
`
`Authors and editors might also find
`confusion in the style difference, Robinson
`says, and gives this advice:
`
`Authors: Check recent
`
`issues of the
`
`publication to which you are submit-
`ting your article to see what convenv
`tion is being followed and ensure that
`your submission conforms. If you are
`uncertain, check with the editorial
`office. Keep this convention in mind
`when you are checking proofs.
`
`Editors: Be alert to the possibility that
`an author from another country may
`be using a different decimal notation;
`if in doubt as to intended meaning,
`seek clarification.
`
`the
`Elise Langdon—Neuner, editor of
`European Medical Writers Association
`journal The Write Smfl. mentions another
`difficulty caused by the style difference.
`She edits reports and manuscripts written
`in English by Austrian scientists and says
`that it can be difficult to ensure that all the
`
`decimal commas in the reports are changed
`to decimal points. “You can be driven mad
`by spotting the commas and changing them
`to points," Langdon-Neuner says. “The soft-
`ware for the tables [the scientists] use is set
`
`. and automatic conver—
`.
`up for commas .
`sion is not possible. Therefore, comma—to—
`point changes have to be made by hant ."
`Another
`potential
`problem that
`Langdon-Neuner points out occurs when
`data files are sent
`from Europe to the
`United States. Some computer programs
`used in the United States are config‘
`ured to interpret commas within numbers
`{European decimals) as thousands markers
`and will convert the decimals to thousands
`
`numbers. For example, some programs will
`convert 1,34 to 1,340 instead of 1.34 as was
`intended. So before sending data to the
`United States, European scientists must
`make sure that all the decimal commas are
`
`changed to points.
`Although the style difference may be a
`bit of a nuisance sometimes.
`it does not
`appear to cause any major problems. "For
`most numbers,
`the convention can be
`discerned from the context or from other
`
`values in the same paper," Robinson says.
`But authors, editors, and readers should be
`aware of the difference. In!
`
`Reference
`l. Cajori F. A history of lrutthematical notations.
`Chicago: The Open Court Publishing Company;
`[923. p 314-335.
`
`Science Editor 0 March — April 2008 ' Vol 3] 0 No 2 I 43
`
`UNITED THERAPEUTICS, EX. 2116
`
`WATSON LABORATORIES V. UNITED THERAPEUTICS, |PR2017-01622
`
`Page 2 of 2
`
`