Exploring the Meaning of Entropy: A Scientific Perspective

  • Thread starter magicfountain
  • Start date
In summary, dE=T*dS holds as a definition of temperature because it comes from the first law of thermodynamics and the definition of entropy. It is a measure of the change in entropy, dS, between two objects. The change in entropy is related to the change in energy, dQ, by the equation dQ=TdS. The coefficients B1 and B2 in dE=T*dS are functions of the energy and can change in value as the objects change internal energy.
  • #1
magicfountain
28
0
dE=T*dS
I don't understand this equation. Does anybody care to show me why that holds?
 
Physics news on Phys.org
  • #2
The equation is not a law so much as a definition (of temperature). Give me some idea of your context as I could give you chapters of books worth of exposition on the subject.

You can start with the idea of Lagrange multipliers and optimization of a quantity subject to constraints. Apply the principle of maximum entropy subject to the constraint that the average energy of a system be fixed to some value E and you get via the method of Lagrange multipliers this equation as a definition of temperature. The flow of heat from higher temperature body to lower temperature body is a behavior increasing entropy which links up the physical idea of temperature to the mathematical definition via Lagrange multipliers.

That is a starting point and you have to go through a lengthy exposition to get to this equation...that is if you want to understand it fully. How much understanding do you require?
 
  • #3
i thought this was the definition of entropy.
how do you assign a numerical value to S (or how do you define it) then?
 
  • #4
Professor Jambaugh did ask for the context.

I would add a further question.

What do you understand the symbols in your equation to stand for and do you understand what they are?
 
  • #5
magicfountain said:
dE=T*dS
I don't understand this equation. Does anybody care to show me why that holds?
It comes from the first law of thermodynamics and the definition of entropy.

The change in entropy, dS is defined as the reversible heat flow, dQ, divided by temperature, T. So dQ = TdS.

From the first law, Q = ΔU + W where W is the work done by the system. I am not sure what dE is supposed to represent. If dE refers to the incremental change in internal energy (ie.: dU) then dE=TdS applies if δW=0.

AM
 
  • #6
OK, try understanding it this way. Firstly take two objects and let them touch. Randomized energy will be exchanged between them. A small change in the energy of one will effect an opposite change in the energy of the other but the total energy will not change.

dE1 + dE2 = 0

Now assume there is an entropy for each object which is a function of the amount of internal randomized energy. S = S(E) and that this entropy is a somewhat additive measure of how randomized a system is. With the small changes in energy there will be a small change in each entropy and we add them:
dS = dS1 + dS2 = B1 dE1 + B2 dE2
where B1 and B2 are the rate coefficients expressing B = dS/dE. These coefficients are again functions of the energy and can change in value as the objects change internal energy.

Substitute in the zero net total energy change and you get:
dS = (B1 - B2)dE1
Assume the 2nd law, that the entropy will grow due to randomization until it reaches a maximum value at which point assuming entropy is a smooth function of energy, you must have the two coefficients equal B1 = B2.

Note that if B1 > B2 increasing the energy E1 increases entropy and vice versa. So we see that heat flows towards the object with the bigger B value. A bit more analysis shows us that the quantity T = 1/B such that:
dE = 1/B dS = T dS
is more useful in that for most objects T will increase with increasing energy.
So we can restate the 2nd law as Heat between two objects flows toward the larger B or the smaller T. In short heat flows from hotter to colder objects.

[edit: Cut out some further less relevant ramblings.]
 
  • #7
thanks very much jambaugh
although the idea of assuming
"there is an entropy for each object which is a function of the amount of internal randomized energy. S = S(E) and that this entropy is a somewhat additive measure of how randomized a system is." seems tough to swallow, your explanation was very helpful and clear and gave me a good notion of what entropy is behind being a measure of "chaos".
 
  • #8
I just posted a specific description of the meaning of entropy in another thread:
https://www.physicsforums.com/showthread.php?t=645089

I hate texts and professors who say entropy is a measure of "chaos" or of "disorder". It is not! It is a measure of ignorance.

But that sounds like it is observer dependent. (and yet "disorder" IS in the eye of the beholder!) It isn't because it is ignorance due to the degree to which the actual system is being physically constrained. The key virtue of science is that "knowledge" has operational meaning in the form of specific measurements made and constraints imposed. We don't "know" because we have a channel to the mystic realm. We know because we physical observe or have physically imposed constraints. Knowledge becomes a physical process. Thus Entropy as an index of relative knowledge (actually ignorance) has perfectly good physical meaning.
 

FAQ: Exploring the Meaning of Entropy: A Scientific Perspective

What is the significance of the equation dE=T*dS in science?

The equation dE=T*dS is known as the Gibbs' equation or the Gibbs' fundamental equation. It is a fundamental equation in thermodynamics that relates the change in energy (dE) of a system to the change in entropy (dS) of the system at a constant temperature (T). This equation has important applications in various fields of science, including chemistry, physics, and engineering.

Where does the equation dE=T*dS come from?

The equation dE=T*dS was first proposed by the American physicist Josiah Willard Gibbs in the late 19th century. Gibbs was a pioneer in the field of thermodynamics and his work laid the foundations for the study of chemical and physical systems at the molecular level.

What does the dE=T*dS equation represent?

The equation dE=T*dS represents the relationship between energy and entropy in a system. It states that the change in energy of a system is equal to the product of the temperature and the change in entropy of the system. This equation is based on the second law of thermodynamics, which states that the total entropy of a closed system will always increase over time.

What are the applications of the dE=T*dS equation?

The dE=T*dS equation has numerous applications in science and engineering. It is used to calculate the change in energy of a system during a chemical reaction, to predict the direction of spontaneous processes, and to understand the behavior of gases and liquids at different temperatures. It is also used in the design and optimization of various industrial processes.

How is the dE=T*dS equation related to entropy?

The dE=T*dS equation is directly related to entropy, which is a measure of the disorder or randomness in a system. According to the equation, the change in energy of a system is directly proportional to the change in entropy at a constant temperature. This means that as the entropy of a system increases, the energy of the system also increases. This relationship is crucial in understanding the behavior of thermodynamic systems.

Back
Top