• 1 Post
  • 138 Comments
Joined 1 year ago
cake
Cake day: June 9th, 2023

help-circle

  • “A computer can never be held accountable, therefore a computer must never make a management decision.”

    Even more importantly when it comes to assessing properly, machine learning, now referred to as AI, has been continuealy shown to not just repeat the biases in its training data, but to significantly exaggerate them.

    Given how significantly and explicitly race has been used to determine and guide so much property and neighborhood development in the training data, I do not look forward to seeing a system that is not only more racist than a post war city council choosing where to build new moterways but which is sold and treated as infallible by the humans operating and litigating it.

    Given the deaths and disaster created by the Horizon Post Office Scandel, I also very much do not look forward to the widespread adoption of software which is inherently and provablly far less accurate, reliable, and auditable than the Horizon software. At least that could only ruin your life if you were a Postmaster and not just any member of the general public who isn’t rich enough to have your affairs handled by a human.

    But hey, on the bright side, if Horizon set UK legal precedent than any person or property agent is fully and unequivocally legally liable for the output of any software they use, after the first few are found guilty for things the procedural text generator they used wrote people might decide its not worth the risk.




  • Generally the term Markov chain is used to discribe a model with a few dozen weights, while the large in large language model refers to having millions or billions of weights, but the fundamental principle of operation is exactly the same, they just differ in scale.

    Word Embeddings are when you associate a mathematical vector to the word as a way of grouping similar words are weighted together, I don’t think that anyone would argue that the general public can even solve a mathematical matrix, much less that they can only comprehend a stool based on going down a row in a matrix to get the mathematical similarity between a stool, a chair, a bench, a floor, and a cat.

    Subtracting vectors from each other can give you a lot of things, but not the actual meaning of the concept represented by a word.


  • To note the obvious, an large language model is by definition at its core a mathematical formula and a massive collection of values from zero to one which when combined give a weighted average of the percentage that word B follows word A crossed with another weighted average word cloud given as the input ‘context’.

    A nuron in machine learning terms is a matrix (ie table) of numbers between zero and 1 by contrast a single human nuron is a biomechanical machine with literally hundreds of trillions of moving parts that darfs any machine humanity has ever built in terms of complexity. This is just a single one of the 86 billion nurons in an average human brain.

    LLM’s and organic brains are completely different and in both design, complexity, and function, and to treat them as closely related much less synonymous betrays a complete lack of understanding of how one or both of them fundamentally functions.

    We do not teach a kindergartner how to write by having them read for thousands of years until they recognize the exact mathematical odds that string of letters B comes after string A, and is followed by string C x percent of the time. Indeed humans don’t naturally compose sentences one word at a time starting from the beginning, instead staring with the key concepts they wish to express and then filling in the phrasing and grammar.

    We also would not expect that increasing from hundreds of years of reading text to thousands would improve things, and the fact that this is the primary way we’ve seen progress in LLMs in the last half decade is yet another example of why animal learning and a word cloud are very different things.

    For us a word actually correlates to a concept of what that word represents. They might make mistakes and missunderstand what concept a given word maps to in a given language, but we do generally expect it to correlate to something. To us a chair is a object made to sit down on, and not just the string of letters that comes after the word the in .0021798 percent of cases weighted against the .0092814 percent of cases related to the collection of strings that are being used as the ‘context’.

    Do I believe there is something intrinsically impossible for a mathematical program to replicate about human thought, probably not. But this this not that, and is nowhere close to that on a fundamental level. It’s comparing apples to airplanes and saying that soon this apple will inevitably take anyone it touches to Paris because their both objects you can touch.


  • Like say, treating a program that shows you the next most likely word to follow the previous one on the internet like it is capable of understanding a sentence beyond this is the most likely string of words to follow the given input on the internet. Boy it sure is a good thing no one would ever do something so brainless as that in the current wave of hype.

    It’s also definitely becuse autocompletes have made massive progress recently, and not just because we’ve fed simpler and simpler transformers more and more data to the point we’ve run out of new text on the internet to feed them. We definitely shouldn’t expect that the field as a whole should be valued what it was say back in 2018, when there were about the same number of practical uses and the foucus was on better programs instead of just throwing more training data at it and calling that progress that will continue to grow rapidly even though the amount of said data is very much finite.


  • Except when it comes to LLM, the fact that the technology fundamentally operates by probabilisticly stringing together the next most likely word to appear in the sentence based on the frequency said words appeared in the training data is a fundamental limitation of the technology.

    So long as a model has no regard for the actual you know, meaning of the word, it definitionally cannot create a truly meaningful sentence. Instead, in order to get a coherent output the system must be fed training data that closely mirrors the context, this is why groups like OpenAi have been met with so much success by simplifying the algorithm, but progressively scrapping more and more of the internet into said systems.

    I would argue that a similar inherent technological limitation also applies to image generation, and until a generative model can both model a four dimensional space and conceptually understand everything it has created in that space a generated image can only be as meaningful as the parts of the work the tens of thousands of people who do those things effortlessly it has regurgitated.

    This is not required to create images that can pass as human made, but it is required to create ones that are truely meaningful on their own merits and not just the merits of the material it was created from, and nothing I have seen said by experts in the field indicates that we have found even a theoretical pathway to get there from here, much less that we are inevitably progressing on that path.

    Mathematical models will almost certainly get closer to mimicking the desired parts of the data they were trained on with further instruction, but it is important to understand that is not a pathway to any actual conceptual understanding of the subject.



  • Not really trying to argue, just trying to help explain the high current DC battery systems I have experience with and to someone how seems to have some conceptual understanding of what individual components do, but not how and why they are used or where the limitations come from.

    Them being confidently incorrect doesn’t help of course. :)

    That being said you haven’t really given me much to work off of as to where these misconceptions are coming from beyond a journalist confusing applications for batteries and capacitors and this really seems to be going nowhere, so bye.


  • Again, there are no capacitors car side to be produceing thermal load in the first place during dc fast charging in the first place, and that thermal load is not the primary barrier to how much current can go into the battery without degradation anyway. After all, if it was we would just upscale the cars heat pump and be charged in five minutes.

    Car charging is not coordinated to the point where they all plug in within a few seconds, and if it was a few second randomizer on when eqch timer actually starts charging would accomplish the exact same effect without hundreds of millions to billions of dollars in new grid scale capacitors and inverters.

    This is also unlikely to become a significant problem because a lot of the grid is moving to battery backed solar and wind, where the limit is price per megawatt hour and as such said batteries can provide far more current than the grid could consume. You might be limited by inverter capacity, but storage capacitors are also fundamentally a DC technology so you would need them anyway.

    This may turn out to have benefits for electronics that rely on already specialized supercapacitors, but it can by definition not have any impact on processes that are not currently limited in any way by capacitor technology like battery bulk charge current, the thing that actually limits how fast a car can fast charge.


  • Obviously nearly every electrical circuit board uses capacitors in some respect, especially for filtering and smoothing, but it is extremely rare for them to be used for bulk energy storage outside of things like adjusting power factor.

    Given we are talking about charging times, which are primarily limited by the batteries charge current vs degradation curve and not at all by the various small capacitors in the charger’s electronics, there is fundamentally no effect on charge times unless you are replacing the energy storage medium itself with supercapacitors.

    We can already supply enough dc power to charge an EV battery at its maximum designed curve via dc fast charging stations, which involve some contractors and shunts but actually don’t even involve any size of capacitors at all in the car itself on the HV side.


  • No, it doesn’t effect devices of all sizes, only devices that might use this specific bulky capacitor, all other devices will show exactly zero improvement because there is no real point to mixing capacitors in with a large battery. Being able to quickly get three minutes of charge per whole hour of battery capacity you replace with capacitors just isn’t that useful because you might as well just stay plugged in for an extra few minutes and get the same charge plus that extra hour before needing to find a charger at all.

    As for EV’s the problem is even more pointless, as being able to go a half mile the street from a charger massive enough that it can output a small power plants worth of electricity is similarly to specialized of a use case to be worth the loss of range and greater degradation of the rest of the battery.


  • The researchers who wrote the paper only mentioned possibly applying the tech to very small things like wearables and Iot applications where a large capacitor might be relevant. It’s the journalist summarizing it that makes the wild claims about phones and cars, which don’t tend to use capacitors for a bunch of reasons, not least of which is that they tend to be physically twenty times larger than a given battery of the same capacity.

    If people are able to deal with batteries anywhere near that large, then I’d imagine most of them would choose twenty times the battery life/ range over being able to charge fast enough overload a wall outlet/ small power plant.


  • While I think in this case they won’t have an effect because no Amarican company is even trying to compete in the space, I feel like claiming “history says tarrifs rarely work” is pretty misleading. The high tarrifs caused by the US generating nearly all federal income by tarrifs in the 17 and 18 hundreds are after all widely credited with being the reason the northern US went from being a minor agricultural nation dependent entirely on european industrial goods to becoming one of the largest industrialized nations so quickly.

    Indeed that was why the WTO blocking third world nations from putting tarrifs on western goods was so heavily criticized by the left a few decades ago, before China proved you could do it without said tarrifs so long as your competitors were greedy enough to outsource their industry to you.


  • While the paper demonstrated strong diminishing returns in adding more data to modern neural networks in terms of image classifers, the video host is explaining how the same may effect apply to any nureal network based system with modern transformers.

    While there are technically methods of generative AI that don’t use a neural network, they haven’t made much progress in recent decades and arn’t what most people mean when they hear or say generative AI, and as such I would say the title is accurate enough for a video meant for a general audience, though “Is there a fundamental limit to modern neural networks” might be more technically correct.



  • I can’t imagine any sort of verification system not being completely overrun by bots/people on fiver/ mechanical turk immediately unless you tied it to meatspace IDs in an know your customer sort of way, in which case you would definitely need a central organization to do said verification, which eliminates any possible need for a blockchain as said organization can just use a faster, far cheaper, and most importantly for this application editable database.

    More to the point, no one doubts that an article published by one organization was secretly published by another, but rather that they secretly used AI in the writing process, which also negates the system because that organization is never going to tell you which articles are done by AI, and any sort of reporting system for the entire organization or a specific author is just going to be immediately and constantly used to review bomb.


  • How could that help at all? Seeing as the blockchain would have no way of telling the difference between human and Ai text, and if you could find a way to automatically verify that in way way that was so efficient you could expect all the text uploaded to the internet you could just run that program locally and not be beholden to people paying a fee to post anything to the internet.