595
temperature
(mander.xyz)
A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.
Rules
This is a science community. We use the Dawkins definition of meme.
Can confirm. Moved from the US to Canada and maybe a year of using Celcius revealed to me just how fucking stupid and convoluted Fahrenheit is. My dad spent three weeks out here and started using Celcius on his phone. Now I only use Fahrenheit when dealing with fevers or temping cases of suspiciously overripe produce.
Fellow Americans. Celcius is superior and more intuitive for those who take a moment to adjust to it. It is okay to accept this as fact without developing an inferiority complex. USA not always #1. USA quite often not #1 and that is okay. It is okay for USA to not be #1 without developing an inferiority complex.
Fahrenheit is European.
*was
I use it and I am not European.
Fahrenheit has a fine granularity that is lost in cold climates. It’s why the Bahamas/Belize use it as well.
Well you know that you can use the decimals?
How is - 40.000001°F more fine than - 40.00000000001°C?
23°C is a nice room temperature.
18°C is a bit chilly but still a comfortable temperature.
If you want to go for a finer destinction then we cann say 18.5°C is warmer but I personally can't feel the difference.
Our bodies are mostly water why not use a system that reflects this?
The universe is mostly empty space with an average temperature of like... 4 Kelvin or some shit. Why not use a system that reflects that? Oh, we do? Right. Celsius is Kelvin + 273.15.
Are you made of mostly empty space? Your response does leave me questioning. Please aknowledge that you are made of 64% water and not 4°k nothing.
I mean, yeah, we all are. That's how atoms work.
alternatively, yeah, mostly between his ears.
As a matter of fact...
Plese do not use Kelvin with a degree symbol. There ist no "degree Kelvin".
Please make sure you are right before you correct someone https://www.techtarget.com/whatis/definition/kelvin-K
I don't know why "techtarget" would be a credible source on Physics questions, but the SI convention, which is, according to Wikipedia, the "only system of measurement with an official status in nearly every country in the world, employed in science, technology, industry, and everyday commerce", poses that "kelvin is never referred to nor written as a degree."
But I also made the mistake to write it as "Kelvin" instead of "kelvin".
...rankine glowers in your general direction...
So then we should use the system that reflects the freezing point and boiling points of water at nice round values such as 0 and 100 then? Sounds like Celsius is the better system
Slightly off topic, but 23°C is a nice room temperature? We have our thermostats at 20°C and I find it quite warm. In the sleeping room we have 18°C and so do I have in my office, which I find quite comfortable. I hate visiting my parents, they always have 22.5°C which I find uncomfortably warm.
Well it's all subjective after all, I'll be happy about chilly 23°C inside when summer comes.
I can feel the difference between 71 and 73 in my house.
At 73, my kids room is uncomfortably hot. At 71, it has a perfect chill for sleeping.
What is your point? That people who use Celsius can't feel the difference between 21.7°C and 22.8°C?
If you're worried about your thermometer, you'll be happy to hear that metric ones usually have finer precision than Fahrenheit ones, since they go in .5°C steps. Since +1°F means +5/9°C, you have less precision!
The point was they need that extra decimal because C isn't good for human temperature sense.
It's not like you are prohibited from using decimals in Fahrenheit. It's that you don't need 3 digits because it works better for people.
And fuck you for making me defend the most ass backwards measurement system on the planet.
It's just an incredibly weak defense. Why is it worse for C to use an extra decimal for these differences? I can just as well argue that C is a more accurate representation, because small differences in temperature are smaller. Just like your argument, this is purely an opinion - until you can show me that not needing the extra decimal is objectively better, or until I can show you that smaller differences being represented as such is objectively better, neither of them holds any weight.
It's the same reason we use abbreviations and contractions when speaking. A trivial simplification is still a simplification.
Why bother with Celcius at all when there is Kelvin. Even Kelvin is arbitrary. Best to use Planck normalized temperature. The scale would be absolute 0 to 100 where 0 is absolute 0 and 100 is 10^32 Kelvin.
So whenever you have to tell someone the temperature outside, you say it's 0.000000000000000000000000015237 Planck
If 3 digits isn't more a tiny bit more cumbersome than 2, then 32 digits is fine too.
We don't have issues with decimals in many places. For example, why are there pennies? Why aren't dollars just scaled up 100? Generally speaking: why don't people immediately shift to the lower unit when talking about e.g. 3.5 miles? If you're correct, those should be simplified too - yet they aren't.
Because Celsius uses a scale that relies on temperatures you're encountering in your everyday life.
Why? That scale is still arbitrarily chosen.
But that's the same reason given for Farenheit!
It's not arbitrary in that it represents the fundamental limits of temperature in the universe. Planck units are fundamental to the nature of the universe rather than based on any arbitrary object.
https://en.m.wikipedia.org/wiki/Planck_units
I would also argue that Fahrenheit is better-suited for everyday life than Kelvin is. Both Celsius and Fahrenheit are objectively closer to temperatures we encounter. Fahrenheit being closer than Celsius is subjective. Do you understand?
There are still a bunch of arbitrary decisions:
All of these are arbitrary decisions you've made when you suggested Planck temperature with a scale from 0 to 100. Do you understand?
Given that you already said you have to use 3 digits to give Celsius the range that matches human temperature sensing, that's not true. 1 degree F is the average threshold that humans can perceive a difference in temperature. It's why thermostats use 3 digits for Celsius but only 2 for Farenheit.
The only reason you say C matches people is because you are used to 21.5 C being a regular indoor temperature. If you grew up with Kelvin that would be 294.5 K. Three digits instead of four.
Doesn't matter. Base 10 would be better so it matches the rest of metric. The decimal place shifts one space but that doesn't change the number of digits needed to represent a temperature.
Zero is absolute zero. You can't have below zero because temperature is a measure of motion.
Linear to match the rest of the metric system.
Dude 71 is way too warm for sleeping, try 64-65 its healthier.
I don't know if my thermostat is just wrong or if the layout of my house makes it inaccurate, but 64-65 in my house is frigid.
Plus we have a baby so 67-68 is really the lowest we could go at night I think.
But I agree, I sleep better in general when the blankets are warm and the house is cold!
Well it's all subjective, I guess. Also depends on where you live.
I would argue it's because of historical usage, familiarity, and resistance to change. Most countries and most people living in hot climates use Celsius.
Save yourself before it's too late.
Do not say anything positive about Fahrenheit in this thread.... the Temperature Scale Inquisition is watching closely for any dissent from the party line.