Lemmy.ca
  • Communities
  • Create Post
  • Create Community
  • heart
    Support Lemmy
  • search
    Search
  • Login
  • Sign Up
RewindAgain@futurology.today to Futurology@futurology.todayEnglish · 2 years ago

Researchers say their open-source AI model LLaVA, is as powerful as GPT-4 Vision, which was released only weeks ago.

llava-vl.github.io

external-link
message-square
15
link
fedilink
36
external-link

Researchers say their open-source AI model LLaVA, is as powerful as GPT-4 Vision, which was released only weeks ago.

llava-vl.github.io

RewindAgain@futurology.today to Futurology@futurology.todayEnglish · 2 years ago
message-square
15
link
fedilink
LLaVA
llava-vl.github.io
external-link
Visual Instruction Tuning
alert-triangle
You must log in or register to comment.
  • Lugh@futurology.todayM
    link
    fedilink
    English
    arrow-up
    17
    ·
    2 years ago

    One of the prevalent doomerist ideas about AI is that big tech will control it all in the future. Yet reality is behaving totally differently. Open-source AI seems only a few months behind big tech.

    • conditional_soup@lemm.ee
      link
      fedilink
      English
      arrow-up
      5
      ·
      2 years ago

      Which is why big tech is currently trying to convince world leadership that only they can be trusted to properly align powerful AI models. It’s almost time to do the ol’ ladder pull.

      • SchizoDenji@lemm.ee
        link
        fedilink
        English
        arrow-up
        2
        ·
        2 years ago

        Corporations would love open source Ai models since it saves them the money that they’d have to pay to someone like openai.

  • sudo22@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    2 years ago

    Can this be easily self hosted?

    • beckerist@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      2 years ago

      https://github.com/haotian-liu/LLaVA#install

    • 🇸‌🇵‌🇪‌🇨‌🇺‌🇱‌🇦‌🇹‌🇪‌🇷‌@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      2
      ·
      2 years ago

      The problem is most of these models need like a terabyte of VRAM… And consumers have about 8-24GB.

      • LoafyLemon@kbin.social
        link
        fedilink
        arrow-up
        7
        ·
        edit-2
        2 years ago

        Old news pal! 😄

        [4/27] Thanks to the community effort, LLaVA-13B with 4-bit quantization allows you to run on a GPU with as few as 12GB VRAM! Try it out here.

        • webghost0101@sopuli.xyz
          link
          fedilink
          English
          arrow-up
          2
          ·
          2 years ago

          12GB of VRAM is still an upgrade away for most people and a 4bit quantized 13B model is barely going to be a tech demo. When open source ai is proclaimed to be near/on par/better then gpt4 they are talking about nothing else than their biggest models in a prime environment.

          • just_another_person@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            2 years ago

            Sure, but not for standard cloud instances that are very affordable for companies wanting to get away from OpenAI.

            • webghost0101@sopuli.xyz
              link
              fedilink
              English
              arrow-up
              1
              ·
              2 years ago

              I usually don’t think much about companies and cloud instances when it comes to Fossai but fair enough.

              For me its all about locally run consumer models. If we cannot archive that it means we will always need to rely on the wims and decisions of others to acces the most transforming technology ever invented.

      • sudo22@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        2 years ago

        Holy shit a terabyte?

        • 🇸‌🇵‌🇪‌🇨‌🇺‌🇱‌🇦‌🇹‌🇪‌🇷‌@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          ·
          2 years ago

          This specific one says it’ll run on 24GB actually. But some are just crazy big.

          • Lugh@futurology.todayM
            link
            fedilink
            English
            arrow-up
            2
            ·
            edit-2
            2 years ago

            There are smaller models that can run on most laptops.

            https://www.maginative.com/article/stability-ai-releases-stable-lm-3b-a-small-high-performance-language-model-for-smart-devices/

            In benchmarks this looks like it is not far off Chat-GPT 3.5.

            • BetaDoggo_@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              2 years ago

              It’s not even close, less than half of 3.5’s 85.5% in ARC. Some larger Open models are competitive in Hellaswag, TruthfulQA and MMLU but ARC is still a major struggle for small models.

              3Bs are kind of pointless right now because the machines with processors capable of running them at a usable speed probably have enough memory to run a 7B anyway.

  • BetaDoggo_@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    1
    ·
    2 years ago

    It’s a solid model but comparing to gpt4 immediately removes any credibility that they had.

Futurology@futurology.today

futurology@futurology.today

Subscribe from Remote Instance

Create a post
You are not logged in. However you can subscribe from another Fediverse account, for example Lemmy or Mastodon. To do this, paste the following into the search field of your instance: [email protected]
Visibility: Public
globe

This community can be federated to other instances and be posted/commented in by their users.

  • 196 users / day
  • 508 users / week
  • 1.78K users / month
  • 6.05K users / 6 months
  • 70 local subscribers
  • 2.56K subscribers
  • 1.76K Posts
  • 10.9K Comments
  • Modlog
  • mods:
  • voidx@futurology.today
  • Lugh@futurology.today
  • Espiritdescali@futurology.today
  • AwesomeLowlander@futurology.today
  • BE: 0.19.11
  • Modlog
  • Legal
  • Instances
  • Docs
  • Code
  • join-lemmy.org