A stopped clock is right specifically twice a day. Any broken clock is right eventually. the only way a clock can be never right is if it works properly and is only desynchronized.
That’s not true. e.g. If a clock loses time as soon as it is started (given power, wound), a time x. Then every day it will be wrong. Now, after n days it will come back around to being correct again. But, if n >> the life of the clock, then no, it will never be correct.
I can think of a few other scenarios where it’s also true that it will never be correct.
But, if n >> the life of the clock, then no, it will never be correct.
After the life of the clock, it will be stopped, and thus right twice per day.
As you said, it may take a very long time to lap the clock, but once you stop drawing distinctions between “never” and “sufficiently infrequent”, you get into the question of acceptable precision. Most people would consider an analog, two-handed clock to be “correct” so long as it is accurate to the minute. That means the threshold of tolerance for a “slow” clock would be the loss of at least one minute per 12 hour period to remain “incorrect”. That means you’ll lap the clock, and it will be correct, every 720 cycles, or about once a year.
If it loses time faster, you’ll lap it faster. If it loses time slower, it will spend more consecutive cycles as “correct” within acceptable tolerance. It’s possible to devise a mechanism which alternates between running fast and slow to ensure that it is actually never correct, but that would have to be built as an accessory mechanism on top of a functioning desynchronized clock in order to ensure that it’s really never.
I’m convinced, the accuracy of the clock matters. Your point that within one minute is on time is fair and as you said converges quickly. Definitely quicker than the life cycle of a regular clock. I’m a convert now.
Wow, broken clock and all that.
Stopped clock, a broken clock may never be right.
A stopped clock is right specifically twice a day. Any broken clock is right eventually. the only way a clock can be never right is if it works properly and is only desynchronized.
If you rip the hands off a clock, it is broken, and it will never be right.
That’s not true. e.g. If a clock loses time as soon as it is started (given power, wound), a time x. Then every day it will be wrong. Now, after n days it will come back around to being correct again. But, if n >> the life of the clock, then no, it will never be correct.
I can think of a few other scenarios where it’s also true that it will never be correct.
After the life of the clock, it will be stopped, and thus right twice per day.
As you said, it may take a very long time to lap the clock, but once you stop drawing distinctions between “never” and “sufficiently infrequent”, you get into the question of acceptable precision. Most people would consider an analog, two-handed clock to be “correct” so long as it is accurate to the minute. That means the threshold of tolerance for a “slow” clock would be the loss of at least one minute per 12 hour period to remain “incorrect”. That means you’ll lap the clock, and it will be correct, every 720 cycles, or about once a year.
If it loses time faster, you’ll lap it faster. If it loses time slower, it will spend more consecutive cycles as “correct” within acceptable tolerance. It’s possible to devise a mechanism which alternates between running fast and slow to ensure that it is actually never correct, but that would have to be built as an accessory mechanism on top of a functioning desynchronized clock in order to ensure that it’s really never.
I’m convinced, the accuracy of the clock matters. Your point that within one minute is on time is fair and as you said converges quickly. Definitely quicker than the life cycle of a regular clock. I’m a convert now.
Oh, uh, I’m not sure what protocol is in this situation. We’re in uncharted Internet-discussion territory here.