• dan1101@lemmy.world
    link
    fedilink
    English
    arrow-up
    99
    arrow-down
    1
    ·
    edit-2
    1 year ago

    As expected, they can’t be trusted. And the more AI evolves, the less likely AI content will be detectable IMO.

    • jocanib@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      32
      arrow-down
      20
      ·
      1 year ago

      It will almost always be detectable if you just read what is written. Especially for academic work. It doesn’t know what a citation is, only what one looks like and where they appear. It can’t summarise a paper accurately. It’s easy to force laughably bad output by just asking the right sort of question.

      The simplest approach for setting homework is to give them the LLM output and get them to check it for errors and omissions. LLMs can’t critique their own work and students probably learn more from chasing down errors than filling a blank sheet of paper for the sake of it.

      • weew@lemmy.ca
        link
        fedilink
        English
        arrow-up
        56
        arrow-down
        2
        ·
        1 year ago

        given how much AI has advanced in the past year alone, saying it will “always” be easy to spot is extremely short sighted.

        • Kara@kbin.social
          link
          fedilink
          arrow-up
          26
          ·
          1 year ago

          People seem to grasp onto weaknesses AI has now and say that they will have them forever, like how text AI lies, and image generation AI can’t draw hands.

          But these AIs are advancing unimaginably quick, 2 years ago generated text was pretty bad, becoming pretty incoherent, and 1 year ago generated images were mostly strange mush.

          • aebrer@kbin.social
            link
            fedilink
            arrow-up
            3
            ·
            1 year ago

            Spot on! Actually people still talk about hands but it’s already been solved with many newer image gen models… The hands they produce look perfectly fine usually these days.

        • Terrasque@infosec.pub
          link
          fedilink
          English
          arrow-up
          6
          arrow-down
          5
          ·
          1 year ago

          Some things are inherent in the way the current LLM’s work. It doesn’t reason, it doesn’t understand, it just predicts the next word out of likely candidates based on the previous words. It can’t look ahead to know if it’s got an answer, and it can’t backtrack to change previous words if it later finds out it’s written itself into a corner. It won’t even know it’s written itself into a corner, it will just continue predicting in the pattern it’s seen, even if it makes little or no sense for a human.

          It just mimics the source data it’s been trained on, following the patterns it’s learned there. At no point does it have any sort of understanding of what it’s saying. In some ways it’s similar to this, where a man learned how enough french words were written to win the national scrabble competition, without any clue what the words actually mean.

          And until we get a new approach to LLM’s, we can only improve it by adding more training data and more layers allowing it to pick out more subtle patterns in larger amounts of data. But with the current approach, you can’t guarantee that what it writes will be correct, or even make sense.

          • nulldev@lemmy.vepta.org
            link
            fedilink
            English
            arrow-up
            6
            arrow-down
            1
            ·
            1 year ago

            it just predicts the next word out of likely candidates based on the previous words

            An entity that can consistently predict the next word of any conversation, book, news article with extremely high accuracy is quite literally a god because it can effectively predict the future. So it is not surprising to me that GPT’s performance is not consistent.

            It won’t even know it’s written itself into a corner

            It many cases it does. For example, if GPT gives you a wrong answer, you can often just send an empty message (single space) and GPT will say something like: “Looks like my previous answer was incorrect, let me try again: blah blah blah”.

            And until we get a new approach to LLM’s, we can only improve it by adding more training data and more layers allowing it to pick out more subtle patterns in larger amounts of data.

            This says nothing. You are effectively saying: “Until we can find a new approach, we can only expand on the existing approach” which is obvious.

            But new approaches come all the time! Advances in tokenization come all the time. Every week there is a new paper with a new model architecture. We are not stuck in some sort of hole.

            • Terrasque@infosec.pub
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              6
              ·
              1 year ago

              An entity that can consistently predict the next word of any conversation, book, news article with extremely high accuracy is quite literally a god because it can effectively predict the future

              I think you’re reading something there other than what I said. Look, today’s LLM’s ingest a ton of text - more accurately tokens - and builds up statistics of which tokens it sees in that context. So statistically if you see the sentence "A nice cup of " statistically the next word is maybe 48% coffee, 28% tea, 17% water and so on. If earlier in the text it says something about heating a cup of oil, that will have a muuch higher chance. It then picks one of the top tokens at (weighted) random, and then the text (array of tokens) is fed in again into the LLM and a new prediction is made. And so on it continues until you stop the loop (usually from a end token or a keyword you’re looking for). Larger LLM’s are better at spotting more subtle patterns - or more accurate it got more layers of statistics that’s applied - but it still has the fundamental issue of going one token at a time and just going by what’s most likely to be the next token.

              It many cases it does. For example, if GPT gives you a wrong answer, you can often just send an empty message (single space) and GPT will say something like: “Looks like my previous answer was incorrect, let me try again: blah blah blah”.

              Have you tried that when it’s correct too? And in that case you mention it has a clean break and then start anew with token generation, allowing it to go a different path. You can see it more clearly experimenting with local LLM’s that have fewer layers to maintain the illusion.

              This says nothing. You are effectively saying: “Until we can find a new approach, we can only expand on the existing approach” which is obvious.

              But new approaches come all the time! Advances in tokenization come all the time. Every week there is a new paper with a new model architecture. We are not stuck in some sort of hole.

              We’re trying to make a flying machine by improving pogo sticks. No matter how well you design the pogo stick and the spring, it will not be a flying machine.

              • nulldev@lemmy.vepta.org
                link
                fedilink
                English
                arrow-up
                5
                ·
                edit-2
                1 year ago

                The issue here is that you are describing the goal of LLMs, not how they actually work. The goal of an LLM is to pick the next most likely token. However, it cannot achieve this via rudimentary statistics alone because the model simply does not have enough parameters to memorize which token is more likely to go next in all cases. So yes, the model “builds up statistics of which tokens it sees in which contexts” but it does so by building it’s own internal data structures and organization systems which are complete black boxes.

                Also, going “one token at a time” is only a “limitation” because LLMs are not accurate enough. If LLMs were more accurate, then generating “one token at a time” would not be an issue because the LLM would never need to backtrack.

                And this limitation only exists because there isn’t much research into LLMs backtracking yet! For example, you could give LLMs a “backspace” token: https://news.ycombinator.com/item?id=36425375

                Have you tried that when it’s correct too? And in that case you mention it has a clean break and then start anew with token generation, allowing it to go a different path. You can see it more clearly experimenting with local LLM’s that have fewer layers to maintain the illusion.

                If it’s correct, then it gives a variety of responses. The space token effectively just makes it reflect on the conversation.

                We’re trying to make a flying machine by improving pogo sticks. No matter how well you design the pogo stick and the spring, it will not be a flying machine.

                To be clear, I do not believe LLMs are the future. But I do believe that they show us that AI research is on the right track.

                Building a pogo stick is essential to building a flying machine. By building a pogo stick, you learn so much about physics. Over time, you replace the spring with some gunpowder to get a mortar. You shape the gunpowder into a tube to get a model rocket and discover the pendulum rocket fallacy. And finally, instead of gunpowder, you use liquid fuel and you get a rocket that can go into space.

                • Terrasque@infosec.pub
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  arrow-down
                  6
                  ·
                  1 year ago

                  The issue here is that you are describing the goal of LLMs, not how they actually work.

                  No, I am describing how they actually work.

                  it cannot achieve this via rudimentary statistics alone because the model simply does not have enough parameters to memorize which token is more likely to go next in all cases.

                  True, hence the limitations. That would require infinite storage and infinite compute capability.

                  Also, going “one token at a time” is only a “limitation” because LLMs are not accurate enough.

                  No, it’s done because one letter at a time is too slow. Tokens are a “happy” medium tradeoff.

                  The space token effectively just makes it reflect on the conversation.

                  It makes a “break” of the block, which lets it start a new answer instead of continuing on the previous. How it reacts to that depends on the fine tune and filters before the data hits the LLM.

                  To be clear, I do not believe LLMs are the future.

                  I have just said that LLM’s we have today can’t fix the problems with false data and hallucinations, because it’s a core principle of how it operates. It will require a new approach.

                  You could add a rocket engine and wings to a pogo stick, but then it’s no longer a pogo stick but an airplane with a weird landing gear. Today’s LLM’s could give us hints to how to make a better AI, but that would be a different thing than today’s LLM’s. From what has been leaked from OpenAI GPT4 has scaling issues so they use mixture of experts. Just throwing hardware at it is already showing diminishing returns. And we’re learning fascinating new ways of training them, but the inherent problem is the same.

                  For example, if you ask an LLM if it can give an answer to a question, it will have two paths to go down, positive and negative. Note, at the point where it chooses that it doesn’t know how to finish it, it doesn’t look ahead. But it sees for example that 80% of the answers in the texts it’s been trained on starts with a positive, then it will most likely start with “yes” - and when it does that it will continue to generate an answer - often very convincing and plausibly real looking answer, because it already committed to that path.

                  And as for the link about teaching it backspace token, the comments there are already pointing out the issue:

                  It’s interesting that in the examples (Table 3 on page 21), the model uses the backspace token to erase the randomly-added token from the prompt, but it does not seem to ever use the token to correct its own output. I’m curious how frequently the model actually uses this backspace token in practice - and if the answer is “vanishingly rarely”, what is the source of the improved Mauve score and sample diversity they show? Is it just that the different training procedure gives an improvement?

                  For it to use the backspace, wouldn’t it have to predict the wrong token with greater confidence than the corrected token? I would think this would require more examples of a wrong token + correction than the correct token, which seems a bit odd.

                  Almost none of the text it’s trained on has a backspace token, and to finetune it in is tricky since it’s a completely new concept - and remember it’s still doing token for token - so it would have to write a token and then right after find out that it’s more likely to send a backspace token than to continue it. It’s interesting, and LLM’s can pick up on some crazy patterns, but I’m skeptical.

      • Zeth0s@lemmy.world
        link
        fedilink
        English
        arrow-up
        24
        ·
        1 year ago

        This is not entirely correct, in my experience. With the current version pf gtp-4 you might be right, but the initial versions were extremely good. Clearly you have to work with it, you cannot ask for the whole work

          • Zeth0s@lemmy.world
            link
            fedilink
            English
            arrow-up
            8
            arrow-down
            1
            ·
            edit-2
            1 year ago

            I meant initial versions of chatGTP 4. ChatGTP isn’t lying, simply because lying implies a malevolent intent. Gtp-4 has no intent, it just provides an output given an input, that can be either wrong or correct. A model able to provide more correct answers is a more accurate model. Computing accuracy for a LLM is not trivial, but gpt-4 is still a good model. User has to know how to use it, what to expect and how to evaluate the result. If they are unable to do so it’s completely their fault.

            Why are you so pissed of a good nlp model?

      • Asifall@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        ·
        1 year ago

        I think there’s a big difference between being able to identify an AI by talking to it and being able to identify something written by an AI, especially if a human has looked over it for obvious errors.

      • Tyler_Zoro@ttrpg.network
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 year ago

        What you are describing is true of older LLMs. GPT4, it’s less true of. GPT5 or whatever it is they are training now will likely begin to shed these issues.

        The shocking thing that we discovered that lead to all of this is that this sort of LLM continues to scale in capabilities with the quality and size of the training set. AI researchers were convinced that this was not possible until GPT proved that it was.

        So the idea that you can look at the limitations of the current generation of LLM and make blanket statements about the limitations of all future generations is demonstrably flawed.

        • jocanib@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          7
          arrow-down
          5
          ·
          1 year ago

          They cannot be anything other than stochastic parrots because that is all the technology allows them to be. They are not intelligent, they don’t understand the question you ask or the answer they give you, they don’t know what truth is let alone how to determine it. They’re just good at producing answers that sound like a human might have written them. They’re a parlour trick. Hi-tech magic 8balls.

          • Tyler_Zoro@ttrpg.network
            link
            fedilink
            English
            arrow-up
            7
            arrow-down
            3
            ·
            1 year ago

            They cannot be anything other than stochastic parrots because that is all the technology allows them to be.

            Are you referring to humans or AI? I’m not sure you’re wrong about humans…

              • nulldev@lemmy.vepta.org
                link
                fedilink
                English
                arrow-up
                4
                ·
                1 year ago

                Have you even read the article?

                IMO it does not do a good job of disproving that “humans are stochastic parrots”.

                The example with the octopus isn’t really about stochastic parrots. It’s more about how LLMs are not multi-modal.

      • Anomander@kbin.social
        link
        fedilink
        arrow-up
        2
        ·
        1 year ago

        I’m no GPT booster, but I think that the real problem with detectability here

        It will almost always be detectable if you just read what is written. Especially for academic work.

        is that it requires you to know the subject and content already, and to be giving the paper a relatively detailed reading. For a rube reading the paper, trying to learn from it - a lot of GPT content is easily mistaken as legitimate. And it’s getting better. We’re not safe simply assuming that AI today is as good as it will ever get and the clear errors we can detect cannot ever be addressed.

        Penetrating academic writing, for academics, is probably one of the highest barriers of any writing task, AI or not.

        But being dismissive of the threat of AI content because it’s not able to convincingly fake some of the hardest writing that real people do is maybe sidestepping a lot of much more casual writing - that still carries significance and consequence.

  • Candelestine@lemmy.world
    link
    fedilink
    English
    arrow-up
    47
    arrow-down
    2
    ·
    1 year ago

    Clearly the Founding Fathers were not advanced enough to have crafted the US Constitution unaided. It’s only reasonable to imagine that ancient aliens could have landed, given them an AI to assist them, and then departed with nobody the wiser.

    I am certain we can find evidence of this if we dig hard enough.

    • Addition@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      16
      ·
      1 year ago

      James Madison was a fraud! He used ChatGPT instead of writing out his homework like he was supposed to. Smh my head.

    • Tyler_Zoro@ttrpg.network
      link
      fedilink
      English
      arrow-up
      6
      ·
      1 year ago

      Clearly the Founding Fathers were not advanced enough to have crafted the US Constitution unaided.

      In a sense you are correct. They cribbed from lots of the most well known political philosophers at the time. For example, there are direct quotes from Locke in the Declaration and his influence over the Constitution can be felt clearly.

    • livus@kbin.social
      link
      fedilink
      arrow-up
      2
      ·
      edit-2
      1 year ago

      I am certain we can find Georgios Tsoukalos saying there is evidence of this sooner or later. Probably extraterrestrial!

    • Zeth0s@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      ·
      1 year ago

      This is a simple example of a constitution. For a real constitution you might want to consider a more detailed list of amendaments and civil rights. Please refer to a professional constitutional writer if you want to create a constitution for productive purposes

    • Gimly@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 year ago

      GPT-4 is quite willing to help you create the constitution for your own country, well I guess I have to find a name now:

      Certainly! Crafting a constitution for a new country is an exciting and monumental task. The constitution is the core document that dictates the legal obligations and rights of your country’s government and its citizens. Here’s a very basic outline to get us started. Please remember, this is a rudimentary draft and it will need to be adjusted based on the specific needs, culture, history, and goals of your country:


      PREAMBLE

      We, the people of [Your Country’s Name], in order to form a more perfect union, establish justice, ensure domestic tranquility, provide for the common defense, promote the general welfare, and secure the blessings of liberty to ourselves and our posterity, do ordain and establish this Constitution for [Your Country’s Name].

      ARTICLE I: THE LEGISLATURE

      Section 1: All legislative powers herein granted shall be vested in a [name of your legislative body].

      Section 2: [Details on how members are elected, their terms, qualifications, powers and responsibilities]

      Section 3: [Details on how legislation is passed, veto powers, etc.]

      ARTICLE II: THE EXECUTIVE

      Section 1: The executive power shall be vested in a [title of your head of state].

      Section 2: [Details on how the head of state is elected, their term, qualifications, powers and responsibilities]

      Section 3: [Details on how the executive branch operates, and other positions within the executive branch]

      ARTICLE III: THE JUDICIARY

      Section 1: The judicial power of [Your Country’s Name] shall be vested in one Supreme Court, and in such inferior courts as the [Legislative body] may from time to time ordain and establish.

      Section 2: [Details on how judges are appointed, their terms, qualifications, powers and responsibilities]

      ARTICLE IV: THE RIGHTS OF THE PEOPLE

      Section 1: All persons born or naturalized in [Your Country’s Name] are citizens of [Your Country’s Name] and of the state wherein they reside. No state shall make or enforce any law which shall abridge the privileges or immunities of citizens.

      Section 2: [Further details on the rights, freedoms, and responsibilities of the citizens, these could include things like freedom of speech, freedom of religion, right to a fair trial, etc.]

      ARTICLE V: AMENDMENTS

      Section 1: The [Legislative body] whenever two thirds of both Houses deem it necessary, shall propose Amendments to this Constitution, or, on the Application of the Legislatures of two thirds of the several states, shall call a Convention for proposing Amendments…

      ARTICLE VI: GENERAL PROVISIONS

      Section 1: [Any other items that need to be addressed, such as the official language, the national religion (or lack thereof), the national emblem, the process for international treaties, etc.]

      ARTICLE VII: RATIFICATION

      This Constitution shall be considered ratified when agreed upon by a majority of citizens through a free and fair vote.


      Please remember that a constitution is a living document that will grow and change over time as your country evolves. This outline is just a starting point and may not include all the specific sections you need. For example, you might want to add sections addressing human rights, environmental protection, financial regulations, and so on.

      This process should involve broad input from many different stakeholders, including legal experts, and should be reflective of the country’s culture, values, history, and aspirations for the future. Also, please note that the effective implementation of a constitution requires a strong rule of law and mechanisms for holding those in power accountable.

  • dethb0y@lemmy.world
    link
    fedilink
    English
    arrow-up
    30
    arrow-down
    1
    ·
    1 year ago

    Because AI detectors suck and are the modern day equivalent of dowsing rods?

    • jocanib@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      22
      arrow-down
      3
      ·
      edit-2
      1 year ago

      They’re circular. If the text is too predictable it was written by an LLM* but LLMs are designed to regurgitate the next word most commonly used by humans in any given context.

      *AI is a complete misnomer for the hi-tech magic 8ball

      • Zeth0s@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        4
        ·
        edit-2
        1 year ago

        The next most commonly used word would result in a loop of common word. LLMs do not work like that

        • jocanib@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          14
          arrow-down
          2
          ·
          1 year ago

          In context. And that is exactly how they work. It’s just a statistical prediction model with billions of parameters.

          • keegomatic@kbin.social
            link
            fedilink
            arrow-up
            11
            arrow-down
            2
            ·
            1 year ago

            That’s not really how LLMs work. You’re basically describing Markov chains. The statement “It’s just a statistical prediction model with billions of parameters” also applies to the human brain. An LLM is much more of a black box than you’re implying.

          • Zeth0s@lemmy.world
            link
            fedilink
            English
            arrow-up
            5
            ·
            edit-2
            1 year ago

            regurgitate the next word most commonly used by humans in any given context.

            is not what it does. That would create non sensical text (you can try yourself).

            This is a summary of the method, as summarized by gtp-4:


            Sure, here is a detailed description of how text is generated with ChatGPT, which is based on the GPT architecture:

            1. Initial Prompt: The process begins with an input prompt. This could be something like “Tell me about the weather today” or any other string of text.
            1. Tokenization: The input text is broken down into smaller parts, called tokens, which can represent words, parts of words, or punctuation. GPT uses a byte pair encoding (BPE) tokenization, which essentially breaks down text into commonly occurring chunks.
            1. Embedding: Each token is then turned into a vector via an embedding. This vector captures semantic information about the token and serves as the input for the model.
            1. Processing the Input: The GPT model processes the input vectors sequentially with a stack of transformer layers. Each layer applies self-attention and feeds its output into the next layer.
            1. Self-Attention Mechanism: The self-attention mechanism in the Transformer model allows it to weigh the importance of different words when predicting the next word. For example, when trying to predict the last word in the sentence “The cat sat on the ____,” the words “cat” and “on” are likely to have more influence on the prediction than “The”. This weighing is learned during training and allows the model to generate more coherent and contextually appropriate responses.
            1. Output Layer: The output from the final transformer layer for the last input token goes through a linear layer followed by a softmax function, which turns it into a probability distribution over the possible next tokens in the vocabulary. Each possible next token is assigned a probability.
            1. Sampling with Temperature: The next token is chosen based on these probabilities. One common method is to sample from this distribution, which introduces some randomness into the process. The temperature parameter controls the amount of randomness: a higher temperature makes the distribution more uniform and the output more random, while a lower temperature makes the model more likely to choose the highest-probability token.
            1. Decoding: The chosen token is then decoded back into text and appended to the output.
            1. Next Iteration: The process then repeats for the next token: the model takes the output so far (including the newly-generated token), processes it, and generates probabilities for the next token. This continues until a maximum length is reached, or an end-of-sequence token is produced.
            1. Post-Processing: Any necessary post-processing is applied, such as cleaning up tokenization artifacts.

            In this way, the model generates a sequence of tokens, one at a time, based on the input prompt and the tokens it has generated so far. Please note that while this process typically uses sampling with a temperature parameter, other methods like beam search or top-k sampling can also be used to choose the next token. These methods have different trade-offs in terms of computational efficiency, diversity, and quality of output.


            You are missing the key part where the text is tranformed in a vector space of “concepts” where semanticic relationships are represented, that is where the inference happens. The inference is not on words to get the next commonly used word, otherwise it wouldn’t work. And you also missed the final sampling to introduce a randomness in the word selection.

            I don’t understand why are you so upset for a chain of complex mathematical functions that complete and input sentence. Why are you angry?

            • jocanib@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              4
              arrow-down
              3
              ·
              1 year ago

              You’re agreeing with me but using more words.

              I’m more annoyed than upset. This technology is eating resources which are badly needed elsewhere and all we get in return is absolute junk which will infest the literature for decades to come.

              • Zeth0s@lemmy.world
                link
                fedilink
                English
                arrow-up
                6
                ·
                1 year ago

                I am not agreeing with you because “regurgitate the next most commonly world” is not what it does.

                That said, the technology is not doing anything wrong. The people using it are doing it. The technology is a great achievement of human kind, possibly one of the greatest. If people decide to use it to print sh*t is people fault. Quantum mechanics is one of the greatest achievement of human kind, if people decided to use it to kill people, it is a fault of people. Many humans are simply shitty, don’t blame a clever mathematical function and its clever implementation

  • Dohnakun@lemmy.fmhy.mlB
    link
    fedilink
    English
    arrow-up
    16
    ·
    1 year ago

    This article was written to keep people as long on the page as possible. It didn’t get to the point before i left. Someone has a tl;dr?

    • Postcard64@lemmy.world
      link
      fedilink
      English
      arrow-up
      18
      ·
      1 year ago

      Constitution is a text that appears many times on the internet. ChatGPT’s training set probably has multiple copies of it. So it’s likely ChatGPT will generate it. Therefore, the detectors are likely to flag it as AI-generated. That’s what I got from it, but I also found it difficult to parse. Maybe someone can correct me on this.

  • busturn@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    ·
    1 year ago

    I’ve recently checked my years-old essay using one of these AI plagiarism detectors and it said that the essay was 90% AI written. So either it’s all bs or I’m a time travelling AI.

    • 98codes@lemm.ee
      link
      fedilink
      English
      arrow-up
      5
      ·
      1 year ago

      I’m convinced that it’s been trained on top of the essays of middle and high school students that have gone their whole lives without proper education on vocabulary, grammar, and the like. So when asked to evaluate something written properly, it’s flagged as AI.

      Garbage in, garbage out. Same as is ever was.

    • paddirn@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 year ago

      Obviously the US Constitution was written by AI, we’re living in a simulation. Wake up sheeple, the Matrix is real!

  • Captain_Patchy@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    They only know what they have been fed.

    What more likely first/base feeding than the US Constitution’s declarations and it’s amendments?