• 0 Posts
  • 16 Comments
Joined 1 year ago
cake
Cake day: June 13th, 2023

help-circle


  • An interesting criteria, why does going back to edit (instead of correcting itself mid-stream)

    I suppose those would be equivalent, I just haven’t seen it done (at least not properly) - the example you posted earlier with the siblings for example was showing how it could only append more text and not actually produce corrections.

    Couldn’t you perform this test on any animal with a discrete brain?

    Oh, right. Animals do exist. It simply hadn’t occurred to me at that moment, even though there is one right next to me taking a nap. However a lot of them are capable of more rational thought than LLMs are. Even bees can count reasonably well. Anyway, defining human level intelligence is a hard problem. Determining it is even harder, but I still say it’s feasible to say some things aren’t it.

    [Garden path sentences]

    No good. The difference between a good garden path and simple ambiguity is that the ‘most likely’ interpretation when the reader is halfway down the sentence turns out to be ungrammatical or nonsense by the end. The way LLMs work, they don’t like to put words together in an order that they don’t usually occur, even if in the end there’s a way to interpret it to make sense.

    The example it made with the keys is particularly bad because the two meanings are nearly identical anyway.

    Just for fun I’ll try to make one here:

    “After dealing with the asbestos, I was asked to lead paint removal.”

    Might not work, the meaningful interpretation could be too obvious compared to the toxic metal, but it has the right structure.


  • What would convince me that we may be on the right path: Besides huge improvements in reasoning, it would (like I mentioned) need to be able to learn - and not just track previous text, I mean permanently adding or adjusting the weights (or equivalent) of the model.

    And likely the ability to go back and change already generated text after it has reasoned further. Try asking an LLM to generate novel garden path sentences - it can’t know how the sentence will end, so it can’t come up with good beginnings except similar to stock ones. (That said it’s not a skill I personally have either, but humans can do it certainly.)

    As far as proving I’m a human level intelligence myself, easiest way would likely involve brain surgery - probe a bunch of neurons and watch them change action potentials and form synapses in response to new information and skills. But short of that, at the current state of the art I can prove it by stating confidently that Samantha has 1 sister. (Note: that thread was a reply to someone else, but I’m watching the whole article’s comments)


  • In your case that was a motor control issue, not a flaw in reasoning. In the LLM case it’s a pure implementation of a Chinese Room and the “book and pencils” (weights) semi-randomly generate text that causes humans to experience textual pareidolia more often than not.

    It can be useful - that book is very large, and contains a lot of residue of valid information and patterns, but the way it works is not how intelligence works (still an open question of course, but ‘not that way’ is quite clear.)

    This is not to say that true AI is impossible - I believe it is possible, but it will have to be implemented differently. At the very least, it will need the ability to self-modify in real time (learning)