The Most Misunderstood Concept in Physics
566 segments
- This is a video about one of the most important,
yet least understood concepts in all of physics.
It governs everything from molecular collisions
to humongous storms.
From the beginning of the universe
through its entire evolution, to its inevitable end.
It may, in fact, determine the direction of time
and even be the reason that life exists.
To see the confusion around this topic,
you need to ask only one simple question.
What does the Earth get from the sun?
- What does the earth get from sun?
- Well, it's light rays?
- What do we get from the sun? - Heat.
- Warmth.
- Warmth, light.
- Vitamin D, we get vitamin D from-
- We do get vitamin D from the ultraviolet rays.
- Well, a lot of energy.
- What does the earth get from this, energy?
- Yeah, energy.
- Energy.
- Nailed it.
Every day, the earth gets a certain amount of energy
from the sun.
And then how much energy does the earth radiate back
into space relative to that amount
that it gets from the sun?
- Probably not as much, I, you know,
I don't believe it's just radiating right back.
- I'd say less.
- Less.
- Less. - I say less.
- I guess about 70%?
- It is a fraction.
- I'd say 20%.
- Because... - Because we use some of it.
- We use some of the energy. - Mm-hmm.
- We consume a lot, right?
- But the thing about energy is it never really goes away.
You can't really use it up.
- It would have to break even, wouldn't it?
Same amount, yeah.
- You know, cause and effect.
It'd be equal in some ways, right?
- For most of the earth's history,
it should be exactly the same amount of energy
in from the sun as earth radiates into space.
- Wow.
- Because if we didn't do that,
then the earth would get a lot hotter, that'd be a problem.
- That'd be a big problem.
- So, if that is the case...
- Yeah.
- Then what are we really getting from the sun?
- That's a good question.
- Hmm.
- It gives us a nice tan.
- It gives us a nice tan, I love it.
We're getting something special from the sun.
- I don't know, what do we get without the energy?
- But nobody talks about it.
To answer that, we have to go back to a discovery
made two centuries ago.
In the winter of 1813, France was being invaded
by the armies of Austria, Prussia, and Russia.
The son of one of Napoleon's generals
was Sadi Carnot, a 17-year-old student.
On December 29th, he writes a letter to Napoleon
to request to join in the fight.
Napoleon preoccupied in battle, never replies.
but Carnot gets his wish a few months later
when Paris is attacked.
The students defend a chateau just east of the city,
but there're no match for the advancing armies,
and Paris falls after only a day of fighting.
Forced to retreat, Carnot is devastated.
Seven years later, he goes to visit his father
who's fled to Prussia after Napoleon's downfall.
His father was not only a general, but also a physicist.
He wrote an essay on how energy
is most efficiently transferred in mechanical systems.
When his son comes to visit, they talk at length
about the big breakthrough of the time, steam engines.
Steam engines were already being used to power ships,
mine ore, and excavate ports.
And it was clear that the future industrial
and military might of nations
depended on having the best steam engines.
But French designs were falling behind
those of other countries like Britain.
So, Sadi Carnot took it upon himself to figure out why.
At the time, even the best steam engines
only converted around 3% of thermal energy
into useful mechanical work.
If he could improve on that, he could give France
a huge advantage and restore its place in the world.
So he spends the next three years studying heat engines,
and one of his key insights
involves how an ideal heat engine would work,
one with no friction and no losses to the environment.
It looks something like this.
Take two really big metal bars, one hot and one cold.
The engine consists of a chamber filled with air,
where heat can only flow in or out through the bottom.
Inside the chamber is a piston,
which is connected to a flywheel.
The air starts at a temperature just below that
of the hot bar.
So first, the hot bar is brought into contact
with the chamber.
The air inside expands with heat flowing into it
to maintain its temperature.
This pushes the piston up, turning the flywheel.
Next, the hot bar is removed,
but the air in the chamber continues to expand,
except now without heat entering, the temperature decreases.
In the ideal case,
until it is the temperature of the cold bar.
The cold bar is brought into contact with the chamber
and the flywheel pushes the piston down.
And as the air is compressed,
heat is transferred into the cold bar.
The cold bar is removed.
The flywheel compresses the gas further
increasing its temperature until it is just below that
of the hot bar.
Then the hot bar is connected again and the cycle repeats.
Through this process, heat from the hot bar is converted
into the energy of the flywheel.
And what's interesting to note about Carnot's ideal engine
is that it is completely reversible.
If you ran the engine in reverse,
first the air expands lowering the temperature,
then the chamber is brought into contact with the cold bar,
the air expands more, drawing in heat from the cold bar.
Next, the air is compressed, increasing its temperature.
The chamber is placed on top of the hot bar
and the energy of the flywheel is used to return the heat
back into the hot bar.
However many cycles were run in the forward direction,
you could run the same number in reverse, and at the end,
everything would return to its original state
with no additional input of energy required.
So by running an ideal engine, nothing really changes.
You can always undo what you did.
So what is the efficiency of this engine?
Since it's fully reversible,
you might expect the efficiency to be 100%,
but that is not the case.
Each cycle, the energy of the flywheel increases
by the amount of heat flowing
into the chamber from the hot bar,
minus the heat flowing out of the chamber at the cold bar.
So to calculate the efficiency, we divide this energy
by the heat input from the hot bar.
Now the heat in on the hot side is equal to the work done
by the gas on the piston,
and this will always be greater than the work done
by the piston on the gas on the cold side,
which equals the heat out.
And this is because on the hot side,
the hot gas exerts a greater pressure
on the piston than that same gas when cold.
To increase the efficiency of the engine,
you could increase the temperature of the hot side,
or decrease the temperature of the cold side, or both.
Lord Kelvin learns of Carnot's ideal heat engine
and realizes it could form the basis
for an absolute temperature scale.
Imagine that the gas is allowed to expand an extreme amount,
so much that it cools to the point
where all the gas particles effectively stop moving.
Then they would exert no pressure on the piston,
and it would take no work to compress it
on the cold side, so no heat would be lost.
This is the idea of absolute zero,
and it would make for a 100% efficient engine.
Using this absolute temperature scale, the Kelvin scale,
we can replace the amount of heat in and out
with the temperature of the hot and cold side respectively,
because they are directly proportional.
So we can express efficiency like this,
which we can rewrite like this.
What we have learned is that the efficiency
of an ideal heat engine doesn't depend on the materials
or the design of the engine,
but fundamentally on the temperatures
of the hot and cold sides.
To reach 100% efficiency, you'd need infinite temperature
on the hot side or absolute zero on the cold side,
both of which are impossible in practice.
So even with no friction or losses to the environment,
it's impossible to make a heat engine 100% efficient.
And that's because to return the piston
to its original position,
you need to dump heat into the cold bar.
So not all the energy stays in the flywheel.
Now, in Carnot's time, high pressure steam engines
could only reach temperatures up to 160 degrees Celsius.
So their theoretical maximum efficiency was 32%,
but their real efficiency was more like 3%.
That's because real engines experience friction,
dissipate heat to the environment,
and they don't transfer heat at constant temperatures.
So for just as much heat going in,
less energy ends up in the flywheel.
The rest is spread out over the walls of the cylinder,
the axle of the flywheel,
and is radiated out into the environment.
When energy spreads out like this,
it is impossible to get it back.
So this process is irreversible.
The total amount of energy didn't change,
but it became less usable.
Energy is most usable when it is concentrated
and less usable when it's spread out.
Decades later, German physicist, Rudolf Clausius,
studies Carnot's engine, and he comes up with a way
to measure how spread out the energy is.
He calls this quantity, entropy.
When all the energy is concentrated in the hot bar,
that is low entropy,
but as the energy spreads to the surroundings,
the walls of the chamber and the axle
will entropy increases.
This means the same amount of energy is present,
but in this more dispersed form,
it is less available to do work.
In 1865, Clausius summarizes the first two laws
of thermodynamics like this.
First, the energy of the universe is constant.
And second, the entropy of the universe tends to a maximum.
In other words, energy spreads out over time.
The second law is core to so many phenomena in the world.
It's why hot things cool down and cool things heat up,
why gas expands to fill a container,
why you can't have a perpetual motion machine,
because the amount of usable energy in a closed system
is always decreasing.
The most common way to describe entropy is as disorder,
which makes sense because it is associated with things
becoming more mixed, random, and less ordered.
But I think the best way to think about entropy
is as the tendency of energy to spread out.
So why does energy spread out over time?
I mean, most of the laws of physics
work exactly the same way forwards or backwards in time.
So how does this clear time dependence arise?
Well, let's consider two small metal bars,
one hot and one cold.
For this simple model,
we'll consider only eight atoms per bar.
Each atom vibrates according to the number
of energy packets it has.
The more packets, the more it vibrates.
So let's start with seven packets of energy
in the left bar and three in the right.
The number of energy packets in each bar
is what we'll call a state.
First, let's consider just the left bar.
It has seven energy packets,
which are free to move around the lattice.
This happens nonstop.
The energy packets hop randomly from atom to atom
giving different configurations of energy,
but the total energy stays the same the whole time.
Now, let's bring the cold bar back in
with only three packets and touch them together.
The energy packets can now hop around
between both bars creating different configurations.
Each unique configuration is equally likely.
So what happens if we take a snapshot at one instant in time
and see where all the energy packets are?
So stop, look at this.
Now there are nine energy packets in the left bar,
and only one in the right bar.
So heat has flowed from cold to hot.
Shouldn't that be impossible because it decreases entropy?
Well, this is where Ludwig Boltzmann
made an important insight.
Heat flowing from cold to hot is not impossible,
it's just improbable.
There are 91,520 configurations
with nine energy packets in the left bar,
but 627,264 with five energy packets in each bar.
That is the energy is more than six times
as likely to be evenly spread between the bars.
But if you add up all the possibilities,
you find there's still a 10.5% chance that the left bar
ends up with more energy packets than it started.
So, why don't we observe this happening around us?
Well, watch what happens as we increase the number of atoms
to 80 per bar and the energy packets to 100,
with 70 in the left bar and 30 in the right.
There is now only a 0.05% chance that the left solid
ends up hotter than it started.
And this trend continues as we keep scaling up the system.
In everyday solids, there are around 100 trillion,
trillion atoms and even more energy packets.
So heat flowing from cold to hot is just so unlikely
that it never happens.
Think of it like this Rubik's cube.
Right now, it is completely solved,
but I'm gonna close my eyes and make some turns at random.
If I keep doing this, it will get further and further
from being solved.
But how can I be confident that I'm really messing
this cube up?
Well, because there's only one way for it to be solved,
a few ways for it to be almost solved,
and quintillions of ways
for it to be almost entirely random.
Without thought and effort,
every turn moves the Rubik's cube
from a highly unlikely state that of it being solved
to a more likely state, a total mess.
So if the natural tendency of energy is to spread out
and for things to get messier,
then how is it possible to have something
like air conditioning where the cold interior of a house
gets cooler and the hot exterior gets hotter?
Energy is going from cold to hot,
decreasing the entropy of the house.
Well, this decrease in entropy is only possible
by increasing the entropy a greater amount somewhere else.
In this case, at a power plant,
the concentrated chemical energy and coal is being released,
heating up the power plant in its environment,
spreading to the turbine the electric generators,
heating the wires all the way to the house,
and producing waste heat in the fans and compressor.
Whatever decrease in entropy is achieved at the house
is more than paid for by an increase in entropy required
to make that happen.
But if total entropy is constantly increasing
and anything we do only accelerates that increase,
then how is there any structure left on earth?
How are there hot parts separate from cold parts?
How does life exist?
Well, if the earth were a closed system,
the energy would spread out completely,
meaning, all life would cease,
everything would decay and mix,
and eventually, reach the same temperature.
But luckily, earth is not a closed system,
because we have the sun.
What the sun really gives us is a steady stream
of low entropy that is concentrated bundled up energy.
The energy that we get from the sun is more useful
than the energy we give back.
It's more compact, it's more clumped together.
Plants capture this energy and use it to grow
and create sugars.
Then animals eat plants and use that energy
to maintain their bodies and move around.
Bigger animals get their energy
by eating smaller animals and so on.
And each step of the way,
the energy becomes more spread out.
- Okay, interesting.
- Yeah.
- Oh wow, I did not know that.
- There you go.
Ultimately, all the energy that reaches earth from the sun
is converted into thermal energy,
and then it's radiated back into space.
But in fact, it's the same amount.
I know this is a- - You do know this is...
- I'm a PhD physicist.
- Oh, okay, but anyway, so... - I trust you.
The increase in entropy can be seen
in the relative number of photons arriving at
and leaving the earth.
For each photon received from the sun,
20 photons are emitted,
and everything that happens on earth,
plants growing, trees falling, herds stampeding,
hurricanes and tornadoes, people eating,
sleeping, and breathing.
All of it happens in the process of converting fewer,
higher energy photons
into 20 times as many lower energy photons.
Without a source of concentrated energy
and a way to discard the spread out energy,
life on earth would not be possible.
It has even been suggested that life itself
may be a consequence of the second law of thermodynamics.
If the universe tends toward maximum entropy,
then life offers a way to accelerate that natural tendency,
because life is spectacularly good
at converting low entropy into high entropy.
For example, the surface layer of seawater produces
between 30 to 680% more entropy when cyanobacteria
and other organic matter is present than when it's not.
Jeremy England takes this one step further.
He's proposed that if there is a constant stream
of clumped up energy, this could favor structures
that dissipate that energy.
And over time, this results
in better and better energy dissipators,
eventually resulting in life.
Or in his own words,
"You start with a random clump of atoms,
and if you shine light on it for long enough,
it should not be so surprising that you get a plant."
So life on earth survives on the low entropy from the sun,
but then where did the sun get its low entropy?
The answer is the universe.
If we know that the total entropy of the universe
is increasing with time, then it was lower entropy yesterday
and even lower entropy the day before that,
and so on, all the way back to the Big Bang.
So right after the Big Bang,
that is when the entropy was lowest.
This is known as the past hypothesis.
It doesn't explain why the entropy was low,
just that it must have been that way for the universe
to unfold as it has.
But the early universe was hot, dense,
and almost completely uniform.
I mean, everything was mixed and the temperature
was basically the same everywhere,
varying by at most 0.001%.
So how is this low entropy?
Well, the thing we've left out is gravity.
Gravity tends to clump matter together.
So taking gravity into account,
having matter all spread out like this,
would be an extremely unlikely state,
and that is why it's low entropy.
Over time, as the universe expanded and cooled,
matter started to clump together in more dense regions.
And in doing so, enormous amounts of potential energy
were turned into kinetic energy.
And this energy could also be used
like how water flowing downhill can power a turbine.
But as bits of matter started hitting each other,
some of their kinetic energy was converted into heat.
So the amount of useful energy decreased.
Thereby, increasing entropy.
Over time, the useful energy was used.
In doing so, stars, planets, galaxies, and life were formed,
increasing entropy all along.
The universe started with around 10
to the 88 Boltzmann constants worth of entropy.
Nowadays, all the stars in the observable universe
have about 9.5 times 10 to the 80.
The interstellar and intergalactic medium combined
have almost 10 times more,
but still only a fraction of the early universe.
A lot more is contained in neutrinos
and in photons of the cosmic microwave background.
In 1972, Jacob Bekenstein proposed
another source of entropy, black holes.
He suggested that the entropy of a black hole
should be proportional to its surface area.
So as a black hole grows, its entropy increases.
Famous physicists thought the idea was nonsense
and for good reason.
According to classical thermodynamics,
if black holes have entropy,
then they should also have a temperature.
But if they have temperatures, they should emit radiation
and not be black after all.
The person who set out to prove Bekenstein wrong
was Stephen Hawking.
But to his surprise, his results showed that black holes
do emit radiation, now known as Hawking radiation,
and they do have a temperature.
The black hole at the center of the Milky Way
has a temperature of about a hundred trillionth of a Kelvin,
emitting radiation that is far too weak to detect.
So still pretty black.
But Hawking confirmed that black holes have entropy
and Bekenstein was right.
Hawking was able to refine Bekenstein's proposal
and determine just how much entropy they have.
The super massive black hole at the center of the Milky Way
has about 10 to the 91 Boltzmann constants of entropy.
That is 1,000 times as much
as the early observable universe,
and 10 times more than all the other particles combined.
And that is just one black hole.
All black holes together account for 3 times 10
to the 104 Boltzmann constants worth of entropy.
So almost all the entropy of the universe
is tied up in black holes.
That means, the early universe
only had about 0.000000000000003% of the entropy it has now.
So the entropy was low, and everything that happens
in the universe like planetary systems forming,
galaxies merging, asteroids crashing,
stars dying, to life itself flourishing,
all of that can happen because the entropy
of the universe was low and it has been increasing,
and it all happens only in one direction.
We never see an asteroid uncrash
or a planetary system unmix
into the cloud of dust and gas that made it up.
There is a clear difference between going to the past
and the future, and that difference comes from entropy.
The fact that we are going from unlikely
to more likely states is why there is an arrow of time.
This is expected to continue until eventually,
the energy gets spread out so completely
that nothing interesting will ever happen again.
This is the heat death of the universe.
In the distant future,
more than 10 to the 100 years from now,
after the last black hole has evaporated,
the universe will be in its most probable state.
Now, even on large scales, you would not be able to tell
the difference between time moving forwards or backwards,
and the arrow of time itself would disappear.
So it sounds like entropy is this awful thing
that leads us inevitably
towards the dullest outcome imaginable.
But just because maximum entropy has low complexity
does not mean that low entropy has maximum complexity.
It's actually more like this tea and milk.
I mean, holding it like this is not very interesting.
But as I pour the milk in, the two start to mix
and these beautiful patterns emerge.
They arise in an instant and before you know it,
they're gone back to being featureless.
Both low and high entropy are low in complexity.
It's in the middle where complex structures
appear and thrive.
And since that's where we find ourselves,
let's make use of the low entropy we've got while we can.
With the right tools, we can understand just about anything,
from a cup of tea cooling down
to the evolution of the entire universe.
And if you're looking for a free and easy way
to add powerful tools to your arsenal,
then you should check out
this video sponsor, brilliant.org.
With Brilliant, you can master key concepts in everything
from math and data science to programming and physics.
All you need to do is set your goal,
and Brilliant will design the perfect learning path for you,
equipping you with all the tools you need to reach it.
Want to learn how to think like a programmer?
Then Brilliant's latest course, "Thinking in Code"
is a fast and easy way to get there.
Using an intuitive drag and drop editor,
it teaches you what you really need to know,
including essential concepts like nesting and conditionals.
You can start by jumping right in to program a robot
and then learn how to apply your new tools
to your everyday life,
like automating reminders on your phone
or building a bot that filters your matches on a dating app.
What I love about Brilliant is that they connect
what you learn to real world examples.
And because each lesson is hands-on,
you'll build real intuition,
so you can put what you've learned to good use.
To try everything Brilliant has to offer free
for a full 30 days, visit brilliant.org/veritasium.
I will put that link down in the description.
And through that link, the first 200 of you to sign up
will get 20% off Brilliant's annual premium subscription.
So I wanna thank Brilliant for sponsoring this video,
and I wanna thank you for watching.
Ask follow-up questions or revisit key timestamps.
The video discusses the concept of entropy, a fundamental principle in physics that governs everything from molecular collisions to the evolution of the universe. It explains that while the total energy in a closed system remains constant, its usability decreases over time as it spreads out, a process quantified by entropy. The video traces the history of this concept from Sadi Carnot's ideal heat engine to Rudolf Clausius's formulation of the second law of thermodynamics and Ludwig Boltzmann's statistical interpretation of entropy. It highlights that entropy is the reason for the arrow of time, the directionality of processes from order to disorder, and from improbable to more probable states. The video also touches upon the role of low entropy from the sun in enabling life on Earth and speculates that life itself might be a consequence of the second law, accelerating the universe's tendency towards maximum entropy. Finally, it explores the origin of low entropy in the universe at the Big Bang and the significant role of black holes in the universe's total entropy.
Videos recently processed by our community