remember to download llama (leaked ai chat model) while you still can

torrent link on their official github:
https://github.com/facebookresearch/llama/pull/73/files

coomers rejoice, globohomo can suck a fat one

  1. 7 months ago
    Anonymous

    Too bad you need a 4090 to run it and even then it's basically useless out of box

    • 7 months ago
      Anonymous

      >4090
      3060
      >out of box
      wrong box
      https://github.com/oobabooga/text-generation-webui/issues/147#issuecomment-1454987216

      • 7 months ago
        Anonymous

        can I do anything with an old 1080TI? I have 64gb of regular memory of that helps

        • 7 months ago
          Anonymous

          don't believe your GPU will help, but you can still run it in RAM with your CPU, it'll just be much slower.

      • 7 months ago
        Anonymous

        Can I do anything with an old RX580?

    • 7 months ago
      Anonymous

      >need a 4090 to run it
      You don't own a CPU, anon?

    • 7 months ago
      Anonymous

      you can rent that gpu for 10cent hour or use google colab for free.

    • 7 months ago
      Anonymous

      LLaMA-7B runs at great speeds on 8GB cards. Better than any other model that could run on those cards.

      I'm running LLaMA-13B on 16GB of VRAM right now.

    • 7 months ago
      Anonymous

      *Multicore CPU and some RAM

  2. 7 months ago
    Anonymous

    It took me 5 hours to download the 220gb file and 3 hours of waiting for seeders... It's just gpt3

    • 7 months ago
      Anonymous

      >it's just gpt3
      Is that bad?

      • 7 months ago
        Anonymous

        It's continuations of prompts granted I only got the 7B model going. It's pretty decent at few shot learning and gpt3 isn't bad, but ya it's not the chatbot ur looking for

  3. 7 months ago
    Anonymous

    The model is pure garbage. Not worth the bandwidth it takes to download. Never trust the zuck.

    • 7 months ago
      Anonymous

      >he didn't convert it
      bro your repetition penalty?

    • 7 months ago
      Anonymous

      It's a bare LLM. It needs further training before it can be your waifu. Think of it as a "how to brain" rather than "what to brain"

    • 7 months ago
      Anonymous

      It's not garbage at all.
      This is 13b with topk and repitiion_penalty

      • 7 months ago
        Anonymous

        And this is princess smut with 35b.

        How much VRAM needs each models?
        I fucking hate when they don't specify the hardware requirement, data "scientist" my ass

        10gb for 7b in 8bit mode and 16gb for 13b in 8bit mode.
        For 35b in 8bit mode, you need 35gb vram

        • 7 months ago
          Anonymous

          can you run 35B on 2x4090s?

          • 7 months ago
            Anonymous

            Yes, people are doing it with 2x3090 and getting good inference speeds.

            • 7 months ago
              Anonymous

              How? Just SLI it?

        • 7 months ago
          Anonymous

          On my system it's closer to 39GB in total.

      • 7 months ago
        Anonymous

        >topk and repitiion_penalty
        What's that?

  4. 7 months ago
    Anonymous

    >"coomers rejoice, globohomo can suck a fat one"
    >BOT user
    >close to 0 knowledge in pytorch, coding, anything ml
    >confirmed retard

    congratulations, you downloaded llama, now what you gonna do, use it on your koboldai gui? oh you gonna generate some text with the 7B model? can't use cli because all your retarded hands can do is press buttons and thats it?

    wannabe researcher go have a nice day gay
    infecting the ai space with retards since august, 2022.... SD should have never became a thing...

    • 7 months ago
      Anonymous

      oobabooga

    • 7 months ago
      Anonymous

      >chimp noises

  5. 7 months ago
    Anonymous

    https://rentry.org/llama-tard

  6. 7 months ago
    Anonymous

    >while you still can
    >GLP-3 License
    >712 Forks
    ?

    • 7 months ago
      Anonymous

      the github forks don't have the model weights in them. however the torrent basically can't be stopped and the weights are on huggingface now

  7. 7 months ago
    Anonymous

    Good morning Sirs can we use this to create anime girl sex story?

    • 7 months ago
      Anonymous

      Yes but even in the story she will be repulsed by our stench,,, indianbros i don't feel so good:(

  8. 7 months ago
    Anonymous

    How much VRAM needs each models?
    I fucking hate when they don't specify the hardware requirement, data "scientist" my ass

  9. 7 months ago
    sage

    At this point Meta should publish the original checksums so that we know if it's the real model. It appears that PyTorch models can execute arbitrary code, so this might as well be a clever attack..

    >anon downloads
    >ai.exe
    >screen instantly goes dark
    >bobs and vagene

    • 7 months ago
      Anonymous

      The torrent referenced in OP's post has a different infohash than the first torrent, it's also missing the llama.sh file that the first torrent has, have not checked individual file hashes yet

      first torrent b8287ebfa04f879b048d4d4404108cf3e8014352

  10. 7 months ago
    CapByte

    i have it working

  11. 7 months ago
    Anonymous

    I'll wait for improvements where I'll be able to run on 8GB VRAM quite fast and without worries. Thank you very much and see you in a month.

    • 7 months ago
      Anonymous

      Your improvements are here. 7B runs on 8GB cards getting up to 30it/s depending on the card. https://github.com/oobabooga/text-generation-webui/issues/147

  12. 7 months ago
    Anonymous

    This is the best faceberg can do? Oy vey!!!

    • 7 months ago
      Anonymous

      It's over, bro. Big Tech has it all. We are doomed. We'll have to use Chinese Meyyg4n models.

  13. 7 months ago
    Anonymous

    https://news.ycombinator.com/item?id=35026902

  14. 7 months ago
    Anonymous

    What’s “it/s”?

    • 7 months ago
      Anonymous

      Tokens per second. A token is ~4 text characters.

    • 7 months ago
      Anonymous

      pronouns

    • 7 months ago
      Anonymous

      Iterations per second.

  15. 7 months ago
    Anonymous

    How good is compared to GPT-J (6B)?

    • 7 months ago
      Anonymous

      I second this question, this is the big question here, and in practice, not on meme metrics used to make papers look good

  16. 7 months ago
    Anonymous

    Will it run on 1060 3GB?

    • 7 months ago
      Anonymous

      not yet

  17. 7 months ago
    Anonymous

    So what I'm getting right now is that this is basically useless for now unless you specifically train it. Is this any different from anything we already have like chatgpt/bing ai or is this just hype unless you're arsed to train it out of box
    2JTJTJ

    • 7 months ago
      Anonymous

      >unless you're arsed to train it out of box
      how is that not appealing? you can train it how you please make it into a BOT shitposter or just an erotica machine.

      • 7 months ago
        Anonymous

        Training is half the fun

        • 7 months ago
          Anonymous

          it's really not
          >change a bit, click submit
          >As an AI...
          >change a bit, click submit
          >As an AI...
          >change a bit, click submit
          >As an AI...
          >change a bit, click submit
          >Error, too many requests in 1 hour
          it sure works fine when it finally decides to give what you want, but getting there is just fucking pain

      • 7 months ago
        Anonymous

        Meta used 2048 a100s to train the model so good luck training champ

  18. 7 months ago
    Anonymous

    I don't know what I would even do with it so I'll pass

  19. 7 months ago
    Anonymous

    Christ almighty super mario

Your email address will not be published. Required fields are marked *