The catarrhine who invented a perpetual motion machine, by dreaming at night and devouring its own dreams through the day.

  • 20 Posts
  • 1.44K Comments
Joined 7 months ago
cake
Cake day: January 12th, 2024

help-circle
  • Note that, even if we refer to Java, Python, Rust etc. by the same word “language” as we refer to Mandarin, English, Spanish etc., they’re apples and oranges - one set is unlike the other, even if both have some similarities.

    That’s relevant here, for two major reasons:

    • The best approach to handle one is not the best to handle the other.
    • LLMs aren’t useful for both tasks (translating and programming) because both involve “languages”, but because LLMs are good to retrieve information. As such you should see the same benefit even for tasks not involving either programming languages or human languages.

    Regarding the first point, I’ll give you an example. You suggested abstract syntax trees for the internal representation of programming code, right? That might work really well for programming, dunno, but for human languages I bet that it would be worse than the current approach. That’s because, for human languages, what matters the most are the semantic and pragmatic layers, and those are a mess - with the meaning of each word in a given utterance being dictated by the other words there.




  • I’ve seen programmers claiming that it helps them out, too. Mostly to give you an idea on how to tackle a problem, instead of copypasting the solution (as it’ll likely not work).

    My main use of the system is

    1. Probing vocab to find the right word in a given context.
    2. Fancy conjugation/declension table.
    3. Spell-proofing.

    It works better than going to Wiktionary all the time, or staring my work until I happen to find some misspelling (like German das vs. dass, since both are legit words spellcheckers don’t pick it up).

    One thing to watch out for is that the translation will be more often than not tone-deaf, so you’re better off not wasting your time with longer strings unless you’re fine with something really sloppy, or you can provide it more context. The later however takes effort.






  • I know you are, but the argument that an LLM doesn’t understand context is incorrect

    Emphasis mine. I am talking about the textual output. I am not talking about context.

    It’s not human level understanding

    Additionally, your obnoxiously insistent comparison between LLMs and human beings boils down to a red herring.

    Not wasting my time further with you.

    [For others who might be reading this: sorry for the blatantly rude tone but I got little to no patience towards people who distort what others say, like the one above.]




  • Upvoted as it’s unpopular even if I heavily disagree with it.

    Look at the big picture.

    There’s a high chance that whatever is causing that flamewar - be it a specific topic, or user conflicts, or whatever - will pop up again. And again. And again.

    You might enjoy watching two muppets shitposting at each other. Frankly? I do it too. (Sometimes I’m even one of the muppets.) However this gets old really fast and, even if you’re emotionally detached of the whole thing, it reaches a point where you roll your eyes and say “fuck, yet another flamewar. I just want to discuss the topic dammit”.

    Plenty people however do not; and once you allow flamewars, you’re basically telling them “if you don’t want to see this, fuck off”. Some of those will be people who are emotionally invested enough in your community to actually contribute with it, unlike the passing troll stirring trouble.


  • I suppose it’s improper to point and laugh? // I see no reason to respond to bad faith arguments.

    It’s improper, sure, but I do worse. You seriously don’t want proselytise Christian babble in my ear if I’m in a bad mood. It sounds like this:

    [Christian] “God exists because you can’t disprove him”

    [Me] “Yeah, just like you can’t disprove that your mum got syphilis from sharing a cactus dildo with Hitler. Now excuse me it’s Sunday morning and I want to sleep.”



  • It doesn’t need to be filtered into human / AI content. It needs to be filtered into good (true) / bad (false) content. Or a “truth score” for each.

    That isn’t enough because the model isn’t able to reason.

    I’ll give you an example. Suppose that you feed the model with both sentences:

    1. Cats have fur.
    2. Birds have feathers.

    Both sentences are true. And based on vocabulary of both, the model can output the following sentences:

    1. Cats have feathers.
    2. Birds have fur.

    Both are false but the model doesn’t “know” it. All that it knows is that “have” is allowed to go after both “cats” and “birds”, and that both “feathers” and “fur” are allowed to go after “have”.





  • IMO it’s more useful to learn how to identify and reply to fallacies and bad premises in general, than to focus on the ones that Christian proselytism uses.

    For example, the ones in the video are:

    • “Either god created us, or we are here by random chance” - false dichotomy + strawman
    • “God exists because you can’t disprove him” - inversion of the burden of the proof
    • “Objective morality proves god exists” - naturalistic fallacy + bad premise
    • “Everything that exists was created. Therefore god exists” - bad premise
    • “You’re not educated enough” - ad hominem

    Others that you need to look for are:

    • invincible authority (a type of appeal to authority) - X was said by authority, thus X is true. Christians love this crap.
    • fallacy fallacy - X is backed up by a fallacy, so X is false
    • ad populum - lots of suckers believe it, so it’s true