The richter magnitude of an earthquake is determined from the logarithm of the amplitude of waves recorded by seismographs. Adjustments are included for the variation in the distance between the various seismographs and the epicenter of the earthquakes.
How Is the Magnitude of an Earthquake Measured?
The magnitude of an earthquake is one of the most important measurements for seismologists and other scientists to understand the power and size of seismic events. An earthquake’s magnitude is determined based on the amount of energy released in the form of seismic waves during an earthquake, and is usually expressed on a logarithmic scale. It is measured by seismometers, instruments that detect and measure the seismic waves that result from an earthquake.
The most common measure of an earthquake’s magnitude is the Richter scale, developed in 1935 by Charles Richter. It is a logarithmic scale that measures the energy released by an earthquake, with a magnitude of one being the lowest and a magnitude of 9 or higher being the highest. Earthquakes with a magnitude of 5 or higher are considered to be major events, while those with a magnitude of 7.0 or higher are known as great earthquakes.
The Richter scale is a measure of the amount of energy released at the earthquake’s epicenter, but seismologists also use the Moment Magnitude Scale (MMS), which measures the total energy released by an earthquake. This is done by measuring the amount of displacement, or “slip,” that takes place along a fault line. The MMS