• 1 Post
  • 13 Comments
Joined 6 months ago
cake
Cake day: January 3rd, 2024

help-circle


  • You seem to be intentionally missing the point, but to reiterate…

    You shower before entering a pool to wash the dirt from your body off (your cleaning yourself).

    The more of your body covered the less effective that shower is.

    Ideally everyone would be naked in the shower, but there are probably outfits which increasingly render the shower less and less effective (e.g. speedos are better than shorts, etc .).

    It would not surprise me if a Burkina covered so much that the cleaning shower is rendered pointless


  • The shower before a pool is to ensure people aren’t entering the pool coated in dirt (e.g. sweat, hair, dead skin, etc…).

    The chemicals in a pool are designed to bind to that dirt and kill any bacteria introduced.

    There is a limit to the chemicals you can add to a pool (before it hurts humans) and once the amount has activated you need to drain the pool and refill it.

    Swimming pools hold crazy amounts of water which is also really expensive to heat up, so pools want to do that as little as possible.

    Clothing interfers with cleaning your body, so people entering near fully clothed (e.g. like a Burkina) will likely introduce more dirt into the pool.

    That translates into increased costs for swimming pools or pools which maintain the old schedule and just operate unsafely.

    This is all based on owning a hot tub and learning how to maintain it.

    Hopefully this also explains why it doesn’t matter people enter the sea fully clothed




  • Firstly it was just a bit of fun but from memory…

    Twitter was listed as having 2 data centers and a couple dozen satellite offices.

    I forgot the data center estimate, but most of those satelites were tiny. Google gave me the floor area for a couple and they were for 20-60 people (assuming a desk consumes 6m2 and dividing the office area by that).

    Assuming an IT department of 20 for such an office is rediculous but I was trying to overestimate.


  • The Silicon Valley companies massively over hired.

    Using twitter as an example, they used to publicly disclose every site and their entire tech stack.

    I have to write proposals and estimates and when Elon decided to axe half the company of 8000 I was curious…

    I assigned the biggest functional team I could (e.g. just create units of 10 and plan for 2 teams to compete on everything). I assumed a full 20 person IT department at every site, etc… Then I added 20% to my total and then 20% again for management.

    I came up with an organisation of ~1200, Twitter was at 8000.

    I had excluded content moderators and ad sellers because I had no experience in estimating that but it gives a idea of the problem.

    I think the idea was to deny competition people but in reality that kind of staff bloat will hurt the big companies


  • Society is complex, visting a country is different from living there an extended period of time and even then even small geographical distances can result in huge changes in culture.

    For example if you started in London and travelled the M4 to Bristol and carried on through Newport and then Cardiff. You would find dramatic differences in housing costs, religiousness, sports played (e.g. football to rugby), views on public transport, job market, jobs people work, education level, favourite drinks, marriage, etc…

    You could spend 3 months basing yourself in any one of those locations and derive completely different views on what is wrong with the UK.

    Which is why the OP brushed this off as nonsense. It also isn’t uncommon for Americans to go somewhere and suggest it would be miles better if it was exactly like the USA, which is why you get the ad hominem.

    It would be like a British Tourist suggesting they don’t drink enough larger or accusing themof being savages for putting salt in tea




  • I actually researched my list, most the technologies were used internally for years and either publically released after better public alternatives had been adopted or it seems buzz reached me years after Google’s first release. So I am wrong.

    Between 2012-2015 I used to consult on Apache Ivy projects (ideally moving them to Maven and purging the insanity people had written). As a result I would get called in when projects had dependency issues.

    The biggest culprits were Guava/GSon, projects would often choose to use them (because Google) and then would discover a bug that had been fixed in a later patch release (e.g. they used 2.2.1 and 2.2.2 had the fix). However the reason they used 2.2.1 was because a library they needed did. Bumping up the version usually caused things to break.

    The standard solution was to ask’why’ they needed Guava/GSon and everytime you would find they are usually some function found in one of the Apache Commons libraries. So I would pull down the commons library rewrite the bit (often they worked identically)

    Fun side note in 2016-2017 I got called to consult on a lot of Gradle projects to fix the same kind of convoluted bespoke things people did with Apache Ivy. Ivy knew the Gradle ‘feautres’ were a massive headache in 2012 and told you to use Maven for those reasons. Ce La vie.

    We tried using Protobuf in 2008 and it was worse than the Apache Axis for JSON conversion (which feels too harsh to say), similarly I had been using AMQP or Kafka for years and tried gRPC when it was released (google say 2016 but I am sure we tried in 2014) and it was worse in every metric I still don’t understand why it exists.

    I was using Vaadin in 2011 and honestly thought GWT was released in 2012. I had to use it in 2014 and the workflow, compile time and look of GWT is just worse than Vaadin.


  • Plex has been baking in features like that to help you see what is on other streaming channels, etc…

    Personally the whole point of Plex for me was it was a container for my existing DVD/Blu Ray collection, while Plex has added some really cool features. Increasingly they keep resetting the dashboard to try and force engagement with new features, it feels a bit user hostile and I’ve been switching to Jellyfin (same idea but entirely open source and self hosted).

    From a discovery perspective, personally I’ve found good content tends to create its own word of mouth style buzz.

    For example at the moment you can’t go near twitter, reddit, work, BBC News, etc… without someone talking about ‘Mr Bates vs The Post Office’. Recently the risa community kept mentioning Babylon 5 so I picked up all 5 seasons for £20 and watched it through. Similarly the Risa community really seems to love Star Trek prodigy so I’ll probably give that a go at some point.


  • The FAANG companies have an internal kind of elitisim that would make staff less effective.

    If you look at any Google Java library, GWT, GSon, Guava, Gradle, Protobuf, etc… there was a commonly used open source library that existed years before that covered 90% of the functionality.

    The Google staff just don’t think to look outside Google (after if Google hasn’t solved it no chance outsiders have) and so wrote something entirely from scratch.

    Then normally within 6 months the open source library has added the killer new feature. The Google library only persists because people hold FAANG as great “Its by Google so it must be good!” Yet it normally has serious issues/limitations.

    The Google libraries that actually suceeded weren’t owned by Google (E.g. Yahoo wrote Hadoop, Kubernetes got spun away from Google control, etc…).


  • I switched my computer illiterate family members to reduce the effort of helping them and they didn’t notice.

    As a helper…

    There are distributions focussed on the latest and greatest (Arch, Fedora, etc…) and ones aiming for stability (Debian, Ubuntu, etc…). Think of them as groups with different views.

    So Linux Mint is Ubuntu but it has the latest Cinnamon desktop. Ubuntu is Debian but focused on fixed releases and adds ‘snaps’ and includes “non-free” by default.

    People have different views on how the desktop should work. The two big desktops are Gnome and KDE.

    Gnome is like Marmite. Its works completely different to any other desktop and people either love it or loathe it. Its often the distribution default.

    With Windows 10/11 I think Microsoft were trying to steal some of KDE’s best features. By default it looks very much like a Windows desktop but lots of people mod it to look/act like macos. Some people struggle with the options it provides.

    Then there are lots of other desktops, for example Cinnamon takes Gnome and turns it into a normal desktop.

    Personally I would suggest Kubuntu as your first attempt. This is a fairly decent install guide.

    Ubuntu tries to minimise the choices you need to make and the 6 month update cycle keeps it fairly stable.

    Kubuntu is Ubuntu it just makes KDE the default instead of Gnome.