←back to thread

287 points moonka | 2 comments | | HN request time: 0.546s | source
Show context
rqtwteye ◴[] No.43562536[source]
I have been in the workforce for almost 30 years now and I believe that everybody is getting more squeezed so they don’t have the time or energy to do a proper job. The expectation is to get it done as quickly as possible and not do more unless told so.

In SW development in the 90s I had much more time for experimentation to figure things out. In the last years you often have some manager where you basically have to justify every thing you do and always a huge pile of work that never gets smaller. So you just hurry through your tasks.

I think google had it right for a while with their 20% time where people could do wanted to do. As far as I know that’s over.

People need some slack if you want to see good work. They aren’t machines that can run constantly on 100% utilization.

replies(25): >>43562590 #>>43562601 #>>43562738 #>>43562748 #>>43562796 #>>43562875 #>>43562911 #>>43562955 #>>43562996 #>>43563116 #>>43563121 #>>43563253 #>>43563309 #>>43563487 #>>43563727 #>>43563795 #>>43563837 #>>43563965 #>>43563995 #>>43564861 #>>43567850 #>>43569250 #>>43569941 #>>43574512 #>>43579456 #
dwattttt ◴[] No.43562996[source]
> People ... aren’t machines that can run constantly on 100% utilization.

You also can't run machines at 100% utilisation & expect quality results. That's when you see tail latencies blow out, hash maps lose their performance, physical machines wear supra-linearly... The list goes on.

replies(2): >>43563635 #>>43565805 #
dehrmann ◴[] No.43563635[source]
The standard rule for CPU-bound RPC server utilization is 80%. Any less and you could use fewer machines; any more and latency starts to take a hit. This is when you're optimizing for latency. Throughput is different.
replies(1): >>43576086 #
1. pdhborges ◴[] No.43576086[source]
Doesn't this depend on the number of servers, crash rates and recovery times? I wouldn't feel confident running 3 servers running at 80% capacity in ultra low latency scenarios. A single crash would overwhelm the other 2 servers in no time.
replies(1): >>43578627 #
2. dehrmann ◴[] No.43578627[source]
Right; this is only for large pools of servers.