In 1859, scientist Robert Kirchhoff introduced an interesting problem into the world of physics: the question of blackbody radiation. A "blackbody" is basically a black box that absorbs all the radiation that is directed toward it. The amount of energy that it emits is independent of the size or shape of the box; it depends only on temperature.

For decades, physicists worked to figure out the relationship between the temperature of the blackbody and the distribution of the emitted energy along the electromagnetic spectrum. This was of particular interest to theorists because finding the relationship could yield valuable physical constants that could then be applied to other physics concerns. However, there was a more concrete and technical reason to search for a formula relating energy to temperature. Such an equation could be used as a standard for rating the strength of electric lamps.

For this reason, the imperial bureau of standards–the
Physikalisch- Technische Reichsanstalt–took a special interest in
finding the formula. And, in 1896, a young German physicist working there,
Wilhelm Wien, seemed to have stumbled onto an equation that worked.
With the knowledge of the spectral distribution of the energy at
one temperature, Wien's equation would produce the distribution
for any other temperature. It was an experimentally accurate theory,
but Wien had no explanation for *why* his equation worked;
he knew only that it did.

Meanwhile, Planck was hired to take Kirchhoff's old job at the University of Berlin. Planck spent much of the 1890s studying problems of chemical thermodynamics, specifically entropy. His work in this field led him to the puzzle of blackbody radiation, and he set himself the goal of finding a workable theory that would yield Wien's equation.

But just as Planck thought he'd found the answer, a series of experiments proved that Wien's equation was actually incorrect. Rather than assuming his theory was correct and hoping the empirical data would eventually prove him right, Planck chose to trust the experimental results: Wien's theory was wrong, which mean Planck's was, too. So, in 1900, Planck was forced to start all over again.

At this point, Planck took a revolutionary step, although
he didn't realize it at the time. Unable to get the numbers to
work any other way, he made a bold assumption: Planck posited that
energy was emitted by the black box in tiny, finite packets. This
was an unprecedented move, as it had always been assumed that energy came
in an unbroken continuous wave, not in a series of discrete energy
packets. But the assumption led Planck to an equation that worked,
the equation that would make him famous: *E = hv*.

In this equation, E stands for the total energy of the light source, v is the frequency of the light, and h was a mathematical constant that came to be known as "Planck's constant." If Planck was right, then energy could only be emitted in certain units–multiples of hv. Planck called these units "quanta," Latin for "how much." This equation challenged everything that had been previously thought about energy. But no one, not even Planck, realized this at the time.

Planck's equation worked, and by 1908, everyone in the field had accepted it, but even the best physicists of the time failed to see its implications. Like Planck, they considered the quantum assumption to be nothing more than a convenience, a mathematical abstraction with no consequences for the real world.

Despite this oversight, Planck's work was impressive enough
to draw the attention and admiration of his peers. The new equation would,
in itself, have been enough to make Planck's career. Planck's theory
yielded two new universal constants that related mechanical measures
of energy to temperature measures: *h* and *K*.
Planck called *K* "Boltzmann's constant", a gesture
of appreciation to Ludwig Boltzmann, whose theories had led Planck
to his own grand solution. In 1900, the value of h meant little
to physicists, but *K* meant a great deal.

Knowing that such a constant as *K* existed,
physicists had composed the equation *LKT* = pressure
of a standard unit of gas. In this equation, *L* stands
for the number of molecules in a standard unit of gas and *T* stands
for the absolute temperature of the gas. They knew that the number
of molecules and the temperature of a gas were directly related
to the pressure it exerted, but they didn't know how, since the
values of both *L* and *K* were a
mystery.

Thanks to Planck, physicists could finally derive a value
for *L*. And knowing *L* eventually
led to even more discoveries, including a theoretical confirmation
of the charge of a single electron. This was one of the earliest
connections physicists were able to make between electrodynamics
and atomic theory, and bridging the gap between these two fields
had been one of Planck's highest goals.

He wasn't the only one with this goal. As the impact of
Planck's work grew and grew, his peers sat up and took notice.
In 1908, Planck was nominated for the Nobel Prize in physics for
the discovery of his two constants and the *E = hv*
formula itself. But Planck's nomination was voted down, not because
his work wasn't significant enough, but because someone had finally
realized it had even more significant implications. It was pointed
out to the Nobel committee that Planck's equation implied that
energy did not come in a continuum, and, horrified by the thought,
the committee declined to award Planck the prize. Instead, the
1908 Nobel Prize went to Gabriel Lippman, for his work in the new
field of color photography.

Though he lost the prize in 1908 for being too revolutionary, more than ten years later, Planck would finally win his Nobel–not in spite of the revolution his theory was about to cause, but because of it.

Readers' Notes allow users to add their own analysis and insights to our SparkNotes—and to discuss those ideas with one another. Have a novel take or think we left something out? Add a Readers' Note!