出典(authority):フリー百科事典『ウィキペディア(Wikipedia)』「2014/07/14 10:55:19」(JST)
ウィクショナリーにセカンドの項目があります。 |
ウィクショナリーにsecondの項目があります。 |
セカンド (second)
このページは曖昧さ回避のためのページです。一つの語句が複数の意味・職能を有する場合の水先案内のために、異なる用法を一覧にしてあります。お探しの用語に一番近い記事を選んで下さい。このページへリンクしているページを見つけたら、リンクを適切な項目に張り替えて下さい。 |
The second (symbol: s) is the base unit of time in the International System of Units (SI)[1] and is also a unit of time in other systems of measurement (abbreviated s or sec[2]); it is the second division of the hour by sixty, the first division by 60 being the minute.[3] Between AD 1000 (when al-Biruni used seconds) and 1960 the second was defined as 1/86,400 of a mean solar day (that definition still applies in some astronomical and legal contexts).[4][5] Between 1960 and 1967, it was defined in terms of the period of the Earth's orbit around the Sun in 1900,[6] but it is now defined more precisely in atomic terms. Seconds may be measured using mechanical, electric or atomic clocks.
Astronomical observations of the 19th and 20th centuries revealed that the mean solar day is slowly but measurably lengthening and the length of a tropical year is not entirely predictable either; thus the sun–earth motion is no longer considered a suitable basis for definition. With the advent of atomic clocks, it became feasible to define the second based on fundamental properties of nature. Since 1967, the second has been defined to be:
the duration of 9192631770 periods of the radiation corresponding to the transition between the two hyperfine levels of the ground state of the caesium 133 atom.[1]
In 1997, the CIPM affirmed that the preceding definition "refers to a caesium atom at rest at a temperature of 0 K."[1]
SI prefixes are frequently combined with the word second to denote subdivisions of the second, e.g., the millisecond (one thousandth of a second), the microsecond (one millionth of a second), and the nanosecond (one billionth of a second). Though SI prefixes may also be used to form multiples of the second such as kilosecond (one thousand seconds), such units are rarely used in practice. The more common larger non-SI units of time are not formed by powers of ten; instead, the second is multiplied by 60 to form a minute, which is multiplied by 60 to form an hour, which is multiplied by 24 to form a day.
The second is also the base unit of time in the centimetre-gram-second, metre-kilogram-second, metre-tonne-second, and foot-pound-second systems of units.
Under the International System of Units (via the International Committee for Weights and Measures, or CIPM), since 1967 the second has been defined as the duration of 9192631770 periods of the radiation corresponding to the transition between the two hyperfine levels of the ground state of the caesium 133 atom.[1] In 1997 CIPM added that the periods would be defined for a caesium atom at rest, and approaching the theoretical temperature of absolute zero (0 K), and in 1999, it included corrections from ambient radiation.[1] Absolute zero implies no movement, and therefore zero external radiation effects (i.e., zero local electric and magnetic fields).
The second thus defined is consistent with the ephemeris second, which was based on astronomical measurements. (See History below.) The realization of the standard second is described briefly in a special publication from the National Institute of Standards and Technology,[7] and in detail by the National Research Council of Canada.[8]
1 international second is equal to:
The Egyptians subdivided daytime and nighttime into twelve hours each since at least 2000 BC, hence the seasonal variation of their hours. The Hellenistic astronomers Hipparchus (c. 150 BC) and Ptolemy (c. AD 150) subdivided the day sexagesimally and also used a mean hour (1⁄24 day), simple fractions of an hour (1⁄4, 2⁄3, etc.) and time-degrees (1⁄360 day or four modern minutes), but not modern minutes or seconds.[9]
The day was subdivided sexagesimally, that is by 1⁄60, by 1⁄60 of that, by 1⁄60 of that, etc., to at least six places after the sexagesimal point (a precision of better than 2 microseconds) by the Babylonians after 300 BC. For example, six fractional sexagesimal places of a day was used in their specification of the length of the year, although they were unable to measure such a small fraction of a day in real time. As another example, they specified that the mean synodic month was 29;31,50,8,20 days (four fractional sexagesimal positions), which was repeated by Hipparchus and Ptolemy sexagesimally, and is currently the mean synodic month of the Hebrew calendar, though restated as 29 days 12 hours 793 halakim (where 1 hour = 1080 halakim).[10] The Babylonians did not use the hour, but did use a double-hour lasting 120 modern minutes, a time-degree lasting four modern minutes, and a barleycorn lasting 31⁄3 modern seconds (the helek of the modern Hebrew calendar),[11] but did not sexagesimally subdivide these smaller units of time. No sexagesimal unit of the day was ever used as an independent unit of time.
In 1000, the Persian scholar al-Biruni gave the times of the new moons of specific weeks as a number of days, hours, minutes, seconds, thirds, and fourths after noon Sunday.[5] In 1267, the medieval scientist Roger Bacon stated the times of full moons as a number of hours, minutes, seconds, thirds, and fourths (horae, minuta, secunda, tertia, and quarta) after noon on specified calendar dates.[12] Although a third for 1⁄60 of a second remains in some languages, for example Polish (tercja) and Turkish (salise), the modern second is subdivided decimally.
The earliest clocks to display seconds appeared during the last half of the 16th century. The earliest spring-driven timepiece with a second hand which marked seconds is an unsigned clock depicting Orpheus in the Fremersdorf collection, dated between 1560 and 1570.[13]:417–418[14] During the 3rd quarter of the 16th century, Taqi al-Din built a clock with marks every 1/5 minute.[15] In 1579, Jost Bürgi built a clock for William of Hesse that marked seconds.[13]:105 In 1581, Tycho Brahe redesigned clocks that displayed minutes at his observatory so they also displayed seconds. However, they were not yet accurate enough for seconds. In 1587, Tycho complained that his four clocks disagreed by plus or minus four seconds.[13]:104
The second first became accurately measurable with the development of pendulum clocks keeping mean time (as opposed to the apparent time displayed by sundials). In 1644, Marin Mersenne calculated that a pendulum with a length of 39.1 inches (0.994 m) would have a period at one standard gravity of precisely two seconds, one second for a swing forward and one second for the return swing, enabling such a pendulum to tick in precise seconds.
In 1670, London clockmaker William Clement added this seconds pendulum to the original pendulum clock of Christiaan Huygens.[16] From 1670 to 1680, Clement made many improvements to his clock and introduced the longcase or grandfather clock to the public. This clock used an anchor escapement mechanism with a seconds pendulum to display seconds in a small subdial. This mechanism required less power, caused less friction and was accurate enough to measure seconds reliably as one-sixtieth of a minute than the older verge escapement. Within a few years, most British precision clockmakers were producing longcase clocks and other clockmakers soon followed. Thus the second could now be reliably measured.
As a unit of time, the second (meaning the second division by 60 of an hour) entered English in the late 16th century, about a hundred years before it was measured accurately. Those who wrote in Latin, including scientists like Bacon, Tycho and Kepler, used the Latin term secunda with the same meaning as far back as the 1200s.
In 1832, Gauss proposed using the second as the base unit of time in his millimeter-milligram-second system of units. The British Association for the Advancement of Science (BAAS) in 1862 stated that "All men of science are agreed to use the second of mean solar time as the unit of time."[17] BAAS formally proposed the CGS system in 1874, although this system was gradually replaced over the next 70 years by MKS units. Both the CGS and MKS systems used the same second as their base unit of time. MKS was adopted internationally during the 1940s, defining the second as 1/86,400th of a mean solar day.
In 1956, the second was redefined in terms of a year (the period of the Earth's revolution around the Sun) for a particular epoch because, by then, it had become recognized that the Earth's rotation on its own axis was not sufficiently uniform as a standard of time. The Earth's motion was described in Newcomb's Tables of the Sun (1895), which provided a formula for estimating the motion of the Sun relative to the epoch 1900 based on astronomical observations made between 1750 and 1892.[6]
The second was thus defined as:
the fraction 1/31,556,925.9747 of the tropical year for 1900 January 0 at 12 hours ephemeris time.[6]
This definition was ratified by the Eleventh General Conference on Weights and Measures in 1960, which also established the International System of Units.
The tropical year in the 1960 definition was not measured but calculated from a formula describing a mean tropical year that decreased linearly over time, hence the curious reference to a specific instantaneous tropical year. This was in conformity with the ephemeris time scale adopted by the IAU in 1952.[18] This definition brings the observed positions of the celestial bodies into accord with Newtonian dynamical theories of their motion. Specifically, those tables used for most of the 20th century were Newcomb's Tables of the Sun (used from 1900 through 1983) and Brown's Tables of the Moon (used from 1923 through 1983).[6]
Thus, the 1960 SI definition abandoned any explicit relationship between the scientific second and the length of a day, as most people understand the term. With the development of the atomic clock in the early 1960s, it was decided to use atomic time as the basis of the definition of the second, rather than the revolution of the Earth around the Sun.
Following several years of work, Louis Essen from the National Physical Laboratory (Teddington, England) and William Markowitz from the United States Naval Observatory (USNO) determined the relationship between the hyperfine transition frequency of the caesium atom and the ephemeris second.[6][19] Using a common-view measurement method based on the received signals from radio station WWV,[20] they determined the orbital motion of the Moon about the Earth, from which the apparent motion of the Sun could be inferred, in terms of time as measured by an atomic clock. They found that the second of ephemeris time (ET) had the duration of 9,192,631,770 ± 20 cycles of the chosen caesium frequency.[19] As a result, in 1967 the Thirteenth General Conference on Weights and Measures defined the SI second of atomic time as:
the duration of 9,192,631,770 periods of the radiation corresponding to the transition between the two hyperfine levels of the ground state of the caesium-133 atom.[6]
This SI second, referred to atomic time, was later verified to be in agreement, within 1 part in 1010, with the second of ephemeris time as determined from lunar observations.[21] (Nevertheless, this SI second was already, when adopted, a little shorter than the then-current value of the second of mean solar time.[22][23])
During the 1970s it was realized that gravitational time dilation caused the second produced by each atomic clock to differ depending on its altitude. A uniform second was produced by correcting the output of each atomic clock to mean sea level (the rotating geoid), lengthening the second by about 1×10−10. This correction was applied at the beginning of 1977 and formalized in 1980. In relativistic terms, the SI second is defined as the proper time on the rotating geoid.[24]
The definition of the second was later refined at the 1997 meeting of the BIPM to include the statement
This definition refers to a caesium atom at rest at a temperature of 0 K.
The revised definition seems to imply that the ideal atomic clock contains a single caesium atom at rest emitting a single frequency. In practice, however, the definition means that high-precision realizations of the second should compensate for the effects of the ambient temperature (black-body radiation) within which atomic clocks operate, and extrapolate accordingly to the value of the second at a temperature of absolute zero.
Today, the atomic clock operating in the microwave region is challenged by atomic clocks operating in the optical region. To quote Ludlow et al.,[25] “In recent years, optical atomic clocks have become increasingly competitive in performance with their microwave counterparts. The overall accuracy of single-trapped-ion-based optical standards closely approaches that of the state-of-the-art caesium fountain standards. Large ensembles of ultracold alkaline earth atoms have provided impressive clock stability for short averaging times, surpassing that of single-ion-based systems. So far, interrogation of neutral-atom-based optical standards has been carried out primarily in free space, unavoidably including atomic motional effects that typically limit the overall system accuracy. An alternative approach is to explore the ultranarrow optical transitions of atoms held in an optical lattice. The atoms are tightly localized so that Doppler and photon-recoil related effects on the transition frequency are eliminated.”
The NRC[dead link] attaches a "relative uncertainty" of 2.5×10−11 (limited by day-to-day and device-to-device reproducibility) to their atomic clock based upon the 127I2 molecule, and is advocating use of an 88Sr ion trap instead (relative uncertainty due to linewidth of 2.2×10−15). See magneto-optical trap and "Trapped ion optical frequency standards". National Physical Laboratory. Such uncertainties rival that of the NIST-F1 caesium atomic clock in the microwave region, estimated as a few parts in 1016 averaged over a day.[26][27]
SI prefixes are commonly used to measure time less than a second, but rarely for multiples of a second (which is known as metric time). Instead, the non-SI units minutes, hours, days, Julian years, Julian centuries, and Julian millennia are used.
Submultiples | Multiples | |||||
---|---|---|---|---|---|---|
Value | Symbol | Name | Value | Symbol | Name | |
10−1 s | ds | decisecond | 101 s | das | decasecond | |
10−2 s | cs | centisecond | 102 s | hs | hectosecond | |
10−3 s | ms | millisecond | 103 s | ks | kilosecond | |
10−6 s | µs | microsecond | 106 s | Ms | megasecond | |
10−9 s | ns | nanosecond | 109 s | Gs | gigasecond | |
10−12 s | ps | picosecond | 1012 s | Ts | terasecond | |
10−15 s | fs | femtosecond | 1015 s | Ps | petasecond | |
10−18 s | as | attosecond | 1018 s | Es | exasecond | |
10−21 s | zs | zeptosecond | 1021 s | Zs | zettasecond | |
10−24 s | ys | yoctosecond | 1024 s | Ys | yottasecond | |
Common prefixes are in bold |
For specialized purposes, a second may be used as a unit of time in time scales where the precise length differs slightly from the SI definition. One such time scale is UT1, a form of universal time. McCarthy and Seidelmann refrain from stating that the SI second is the legal standard for timekeeping throughout the world, saying only that "over the years UTC [which ticks SI seconds] has become either the basis for legal time of many countries, or accepted as the de facto basis for standard civil time".[28]
Look up second in Wiktionary, the free dictionary. |
Wikimedia Commons has media related to Second. |
|
|
|
|
全文を閲覧するには購読必要です。 To read the full text you will need to subscribe.
リンク元 | 「100Cases 52」「二次」「第二」「quadratic」「秒」 |
拡張検索 | 「secondary progressive multiple sclerosis」「second dimension」 |
.