Meta “programmed it to simply not answer questions,” but it did anyway.

  • rottingleaf@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    3
    ·
    4 months ago

    Why are you writing this to me?

    Do you know what a syllogism is?

    It doesn’t require being certain of the information we’re building it on. Only of existence of such categories.

    Naturally people in Antiquity and Middle Ages who used symbolic logic were even less certain of the actual truths and lies in the world than we are.

    It allows the truth to be subjective, but not the logical constructions. This is a very important trait both then and now.

    The difference between the filter and the data going through it.

    Of course you can’t just feed all the data of all the PoVs and similar cases on something, integrate it into a model and expect your PoV to not clash with its output.

    It’s philosophically the same as why using dialectics is bad for science.

    • conciselyverbose@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      1
      ·
      4 months ago

      A syllogism is a tool for theoretical reasoning that doesn’t actually apply in the real world, because it relies on Boolean possibility spaces. There is never an “all articles by X are correct”, and there is no theoretical possibility that “all articles by X are correct” in the real world. The connections in the real world are literally always probabilistic. In every case. Every time.

      You can’t use formal logic for any real world use case because there are no valid starting assumptions. The only thing logic can ever prove is internal consistency, not fact.

      • rottingleaf@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        4 months ago

        The only thing logic can ever prove is internal consistency, not fact.

        Yes, and being able to build structures with internal consistency would be an advantage.

        Nobody says you can prevent any “AI” oracle from saying things that aren’t true.

        But a tool which would generate a tree of possible logical conclusions from something given in language and then divided into statements on objects with statistical dependencies could be useful.