• 0 Posts
  • 10 Comments
Joined 9 months ago
cake
Cake day: January 31st, 2024

help-circle


  • Can’t speak for the science libraries as I’ve never used em, and I’ll gladly just blindly accept that as truth, but for everything else it’s always a pain in the ass. For being designed to “run on anything” it sure is funny that 90% of the time I download a python app it doesn’t fucking work and requires me to look up and manually setup a specific environment for it. Doesn’t help that the error messages are usually completely random and unrelated to this…

    I always dread when some fucking madman makes the installer for their app in python, knowing it’ll probably fail… God forbid it’s a script that’s supposed to modify something else. Always a good time for reflection upon the choices that led me to this point.

    Even my old scripts I kept around for sentimental value. Half of those don’t work either, and I can’t be bothered to figure out what version I made em for.

    I tried my best to scrub python from my pc out of principle, but as you say, it’s soo common my distro uses it as a dependency, fucking bullshit!



  • I download my music and order it by Artists>Album>Song, basically without exception. Occasionally annoying when a song has multiple authors, because people don’t always write the metadata the same way and it fucks with my music player, but that’s besides the point.

    When I make playlist, I just take a whole album, filter out a few songs if need be and shove it into a given playlist, sometimes I can do that with an entire artist, but it’s not always that easy.

    Another issue with my approach is the odd single song from a random artist that’s really good, but everything else they ever made makes me fall asleep, that’s a really annoying one… Might start making my own fake albums.


  • LANIK2000@lemmy.worldtoScience Memes@mander.xyzHiiiiiii!
    link
    fedilink
    English
    arrow-up
    24
    ·
    4 months ago

    Why does this keep appearing? Cats most definitely do communicate by meowing too. I’d really love it if our village cats didn’t chose the street out my widow a midnight as their debate grounds, shit’s loud af and I always have to climb out the window to break up the fight.


  • LANIK2000@lemmy.worldtoScience Memes@mander.xyzAcademia to Industry
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    4 months ago

    It’s not much of a trick question, if it’s absolutely trivial. It’s cherry picked to show that the AI tries to associate things based on what they look like, not based on the logic and meaning behind them. If you gave the same prompt to a human, they likely wouldn’t even think of the original riddle.

    Even in your example it starts off by doing absolute nonsense and upon you correcting it by spelling out the result, it finally manages, but still presents it in the format of the original riddle.

    You can notice, in my example I intentionally avoid telling it what to do, rather just question the bullshit it made, and instead of thinking “I did something wrong, let’s learn”, it just spits out more garbage with absolute confidence. It doesn’t reason. Like just try regenerating the last answer, but rather ask it why it sent the man back, don’t do any of the work for it, treat it like a child you’re trying to teach something, not a machine you’re guiding towards the correct result.

    And yes, people with memory issues immediately suffer on the inteligence side, their lives a greatly impacted by it and it rarely ends well for them. And no, they are not worthless, I never said that they or AI is worthless, just that “machine learning” in its current state (as in how the technology works), doesn’t get us any closer to AGI. Just like a person with severe memory loss wouldn’t be able to do the kind of work we’d expect from an AGI.


  • LANIK2000@lemmy.worldtoScience Memes@mander.xyzAcademia to Industry
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    4 months ago

    Let’s do the riddle I suggested, because we need something popular in the dataset, but present it with a deviation that makes it stupidly simple yet is unlikely to exist.

    Prompt:

    I have a riddle for you. With a boat, how can a man and a goat get across the river?

    Answer:

    Ah, that’s a classic riddle! The answer is: the man takes the goat across the river first and leaves it on the other side. Then he goes back, picks up the boat, and brings it across to the side where he left the goat. Finally, he leaves the goat there and takes himself across. That way, both the man and the goat safely cross the river without any issues. Do you have any other riddles or questions you’d like to discuss today?

    A normal human wouldn’t be fooled by this and say that they can just go across and maybe ask where the riddle is. They’d be likely confused or expect more. The AI doesn’t because it completely lacks the ability to reason. At least it ends up solved, that’s probably the best response I got when trying to make this point. Let’s continue.

    Follow up prompt:

    Why would he need to take the boat? Wouldn’t he need to have used the boat to get the goat across? And once the boat is there, why would he return to take him self?

    Answer:

    Ah, I see where the confusion lies! In this classic riddle scenario, the man and the goat can actually swim across the river without using the boat. The key is to ensure that either the man or the goat is always present on each side of the river while they are crossing together. Thank you for pointing out that oversight in my previous response! If you have any more riddles or questions to discuss, feel free to share them. How else can I be of assistance today?

    Final prompt:

    Why is that the key? What difference does it make? How can they be both present on both sides while crossing together?

    Final answer:

    I apologize for the confusion in my previous responses. Allow me to clarify the solution to this riddle: The key factor in this scenario is ensuring that either the man or the goat remains on each side of the river at all times to prevent any issues that may arise from leaving them alone together. Here’s how they can successfully cross together without using a boat: 1. The man takes the goat across the river first and leaves it on the other side. 2. The man returns alone and brings himself across to join the goat on that side. 3. The man then picks up the goat and carries it back across to the original side of the river. 4. Finally, he goes back one last time to bring himself across. By following this sequence, both can be present on each side while crossing together, ensuring that there are no conflicts or risks during their journey. Thank you for your patience, and I appreciate your attention to detail in understanding this riddle! If you have any more questions or topics you’d like to discuss, feel free to share them with me. How else can I assist you today?

    I think that’s quite enough, it’s starting to ramble like you said it would (tho much earlier than expected) and unlike the first solution, it doesn’t even end up solved anymore xD I’d argue this is a scenario that should be absolutely trivial and yet the AI is trying to assert information that I didn’t present and continues to fail to apply logic correctly. The only time it knows how to reason is when someone in its dataset already spelled out the reasoning to a certain question. If the logic doesn’t exits in the dataset, it has great difficulty making heads or tails of it.

    And yes, I’d argue memories are indeed absolutely vital to inteligence. If we want cognition, aka the process of acquiring knowledge and understanding, we need it to remember. And if it immediately loses that information or it erodes so quickly, it’s essentially worthless.


  • The black box is the human that reads and outputs text and the analytical prediction machine is the AI. 5 years of development is the human living their life before retuning to continue writing. It is an extreme example, but I’m just tyring to point out that the context of what a person might write can change drastically between individual messages because anything can happened in between, and thus the data is fundamentally flawed for training intelligence, as that step is fully missing, the thought process.

    As to why I called the AI an analytical prediction machine, that’s because that’s essentially what it does. It has analyzed an unholy amount of random text from the internet, meaning conversations/blogs/books and so on, to predict what could follow the text you gave it. It’s why prompt injection is so hard to combat and why if you give it a popular riddle and change it slightly like “with a boat, how can a man and goat get across the river”, it’ll fail spectacularly trying to shove in the original answer somehow. I’d say that’s proof it didn’t learn to understand (cognition), because it can’t use logic to reason about a deviation from the dataset.

    As for memory, we can kind of simulate it with text, but it’s not perfect. If the AI doesn’t write it down, it didn’t happen and thus any thoughts, feelings or mental analysis stops existing upon each generation. The only way it could possibly develop intelligence, is if we made it needlessly ramble and describe everything like a very bad book.

    And thus to reach the beginning of your comment, I don’t belive it’s necessary to posses any cognitive abilities to generate text and in turn I don’t see it as evidence of us getting any closer to AGI.


  • LANIK2000@lemmy.worldtoScience Memes@mander.xyzAcademia to Industry
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    4 months ago

    Thing is we’re not feeding it how humans react to stimulus. For that you’d need it hooked up to a brain directly. It’s too filtered and biased by getting text only, this approach naively ignores things like memory and assumes text messages exist in a vacuum. Throwing a black box into an analytical prediction machine, only works as long as you’re certain it’ll generally throw out the same output with the same input, not if your black box can suddenly experience 5 years of development and emerge a different entity. It’s skipping too many steps to become intelligent, I mean it literally skips the entire process between reading and writing.