Linux server admin, MySQL/TSQL database admin, Python programmer, Linux gaming enthusiast and a forever GM.

  • 2 Posts
  • 195 Comments
Joined 1 year ago
cake
Cake day: June 8th, 2023

help-circle




  • No, but most people do think of Reality as a god.

    But they don’t think He is a god

    These two statements directly contradict eachother. Either you think of Reality as a god, or you do not. If you don’t think reality is a god, then you don’t think “he” is a god (wow, the tautology was weird to type).

    Also, there are many examples of religious people who don’t think reality is a god. Deists, for example, believe that god showed up, made the universe, and left. Under that belief system, the deist god and the universe it created are two entirely separate things.

    There’s also pantheons, where gods exist within their own higher reality with their own set of rules, limitations and powers that interact with our reality. Reality (either ours or the supernatural plane in these belief systems) are clearly separate from the gods operating under these rules.

    People are perfectly capable of worshipping, praying to, and generally being religious towards someone they refuse to believe is a god.

    What’s your opinion on people who do neither? That don’t believe in a god and don’t pray to anything either?













  • So, first of all, thank you for the cogent attempt at responding. We may disagree, but I sincerely respect the effort you put into the comment.

    The specific part that I thought seemed like a pretty big claim was that human brains are “simply” more complex neural networks and that the outputs are based strictly on training data.

    Is it not well established that animals learn and use reward circuitry like the role of dopamine in neuromodulation?

    While true, this is way too reductive to be a one to one comparison with LLMs. Humans have genetic instinct and body-mind connection that isn’t cleanly mappable onto a neural network. For example, biologists are only just now scraping the surface of the link between the brain and the gut microbiome, which plays a much larger role on cognition than previously thought.

    Another example where the brain = neural network model breaks down is the fact that the two hemispheres are much more separated than previously thought. So much so that some neuroscientists are saying that each person has, in effect, 2 different brains with 2 different personalities that communicate via the corpus callosum.

    There’s many more examples I could bring up, but my core point is that the analogy of neural network = brain is just that, a simplistic analogy, on the same level as thinking about gravity only as “the force that pushes you downwards”.

    To say that we fully understand the brain, to the point where we can even make a model of a mosquito’s brain (220,000 neurons), I think is mistaken. I’m not saying we’ll never understand the brain enough to attempt such a thing, I’m just saying that drawing a casual equivalence between mammalian brains and neural networks is woefully inadequate.




  • I’m happy with the Oxford definition: “the ability to acquire and apply knowledge and skills”.

    LLMs don’t have knowledge as they don’t actually understand anything. They are algorithmic response generators that apply scores to tokens, and spit out the highest scoring token considering all previous tokens.

    If asked to answer 10*5, they can’t reason through the math. They can only recognize 10, * and 5 as tokens in the training data that is usually followed by the 50 token. Thus, 50 is the highest scoring token, and is the answer it will choose. Things get more interesting when you ask questions that aren’t in the training data. If it has nothing more direct to copy from, it will regurgitate a sequence of tokens that sounds as close as possible to something in the training data: thus a hallucination.