I don’t consider myself very technical. I’ve never taken a computer science course and don’t know python. I’ve learned some things like Linux, the command line, docker and networking/pfSense because I value my privacy. My point is that anyone can do this, even if you aren’t technical.

I tried both LM Studio and Ollama. I prefer Ollama. Then you download models and use them to have your own private, personal GPT. I access it both on my local machine through the command line but I also installed Open WebUI in a docker container so I can access it on any device on my local network (I don’t expose services to the internet).

Having a private ai/gpt is pretty cool. You can download and test new models. And it is private. Yes, there are ethical concerns about how the model got the training. I’m not minimizing those concerns. But if you want your own AI/GPT assistant, give it a try. I set it up in a couple of hours, and as I said… I’m not even that technical.

  • EonNShadow@pawb.social
    link
    fedilink
    English
    arrow-up
    165
    ·
    1 month ago

    “learned some things like Linux, command line, docker, and networking/pfsense” “I don’t consider myself technical”

    Don’t sell yourself short, I work in IT and have colleagues on our helpdesk who would struggle endlessly with those concepts.

    I hereby dub you a tech person, like it or not, those skills can and do pay the bills.

    • damnthefilibuster@lemmy.world
      link
      fedilink
      English
      arrow-up
      55
      ·
      1 month ago

      Now that you’ve dubbed OP a tech person…

      Hey OP, can you help me fix my printer? It’s only printing “RED RUM RED RUM” for some reason.

    • chagall@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      38
      ·
      1 month ago

      This made me smile. Thank you. The grass is always greener and I sometimes daydream of working in IT instead of healthcare. Maybe someday.

        • GissaMittJobb@lemmy.ml
          link
          fedilink
          English
          arrow-up
          7
          ·
          1 month ago

          Healthcare is pretty rough, I’d be willing to bet that the grass actually is greener in this case.

          • Biezelbob@programming.dev
            link
            fedilink
            English
            arrow-up
            4
            ·
            edit-2
            1 month ago

            I am actually considering switching to healthcare (been a professional programmer)

            I’ve had a burnout: I wish it was due caring for people in need instead of a stupid deadline.

            Besides, you can always do IT as a hobby/for free. Harder with healthcare, except maybe volunteering

            • barsquid@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              1
              ·
              1 month ago

              You’ll be saving lives, yeah, but between dealing with entitled assholes that won’t follow directions and then yell at you because they didn’t.

              It’s maybe easy to burn out in any career. Society has deprioritized individual fulfillment for most of us because it harms the nesting levels of billionaires’ yachts.

    • Scrubbles@poptalk.scrubbles.tech
      link
      fedilink
      English
      arrow-up
      8
      ·
      1 month ago

      I was just talking to a member of my devops team and I was talking about this exact thing and they said “I didn’t know you could attach a GPU to a container”. So, yup, just stay on top of this stuff at home and you’ll do fine

  • DavidGarcia@feddit.nl
    link
    fedilink
    English
    arrow-up
    42
    ·
    1 month ago

    people need to take a step back and realize we have the capability to trap quasi-omnipotent quasi-demons in our personal computers

    yeah they lie a lot and rarely do what you want them to, but that’s just what demons do

    And it’s all powered by some dark crystals created with light magic that slowly poison the planet

    that’s some arcane bullshit

    • Last@reddthat.com
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 month ago

      How long can something like that really last, though? I wish we had a better idea of the timeline, before the quasi-demons start freelancing lol

  • coffee_with_cream@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    25
    ·
    1 month ago

    Uncensored models are so much better, too. chatGPT is like one of those plastic children’s toy hammers vs real models are titanium hammers

    • patrick@lemmy.jackson.dev
      link
      fedilink
      English
      arrow-up
      5
      ·
      1 month ago

      Together.ai has a number of uncensored models too. I’ve found that those are so cheap that it’s not worth trying to self just models unless you really need more privacy.

  • HumanPerson@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    15
    ·
    1 month ago

    Yeah, I like it too. My only issue is ollama’s lack of intel support. I have been looking at issue 1590 on their GitHub. For now I have a 1050ti in a cardboard box PC with other hardware being 10+ years old and a mixed set of RAM totalling 12G. It also has a 100Mbit nic, so I can’t take advantage of full internet speed when downloading models. The worst part is they can support intel, but haven’t merged the solution because of an issue with the windows intel drivers. Linux is fine but I can 't have it. I wasn’t planning to rant, but I already typed it so… enjoy?

    • chagall@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      12
      ·
      edit-2
      1 month ago

      Yeah, I have an NVDIA GPU and it is magic. The best part is when you are using Ollama, open a second terminal window and enter the command, watch -n 0.5 nvidia-smi and you can see your GPU usage go up and down in real-time as you ask the GPT questions. Pretty cool.

      Hopefully they get the ARC folks up and running soon.

  • Goodtoknow@lemmy.ca
    link
    fedilink
    English
    arrow-up
    9
    ·
    1 month ago

    Have you found much practical use for small models yet? I love the idea that even the 1.1B tinyllama model can run on my phone, but haven’t found much real world use for it yet. Llama3 8b feels better, but not much better for even emails as it’s a bit dumb

    • chagall@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      6
      ·
      1 month ago

      I use my phone all the time, but I just use a wireguard VPN to tunnel into my home container of Open WebUI. Then I can interact with my desktop machine using a NVIDIA gpu. I’m currently testing mistral-nemo. It’s pretty great but it gets a bit verbose sometimes.

      • kureta@lemmy.ml
        link
        fedilink
        English
        arrow-up
        9
        ·
        1 month ago

        I am also using open webui. Most LLMs are too verbose for me, so I created a model in open-webui with system prompt “Do not repeat the questions. Avoid giving lists as answers. Do not summarize the answer at the end. If asked a follow-up question, respond with only new information, do not repeat previously stated information.” and named it No Nonsense.

        • chagall@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          3
          ·
          1 month ago

          That’s really smart. I just found out about fabric yesterday and it is helping me with things like what you stated. Prompt engineering is a huge thing.

    • coffee_with_cream@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 month ago

      Imo it’s worthwhile to just run the biggest model available and rent expensive GPU time. It still amounts to very little overall and you get much better results. Project dependent of course

  • dan@upvote.au
    link
    fedilink
    English
    arrow-up
    8
    ·
    1 month ago

    It’s a much smaller scale but I use a Coral TPU with CodeProject AI to detect when people or animals are in front of my house. Works well with Blue Iris (NVR software for security cameras). I like it. That’s all the self-hosted AI I’ve got for now.

  • BlackLaZoR@kbin.run
    link
    fedilink
    arrow-up
    8
    ·
    1 month ago

    I access it both on my local machine through the command line

    You really don’t have to - There’s GPT4ALL designed for normal users with very simple GUI

    Also, with minimal command line knowledge you can install InvokeAI - probably the best UX for image generating AI on the market. Works both on Linux and Windows

    • chagall@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      6
      ·
      1 month ago

      It’s so great that there is so much ongoing development of these types of tools out there. I’m currently using openweb ui as my GUI but I’ll give your suggestion a try next week. I haven’t figured out a use case for stable diffusion except for creating new content for the shitposting community on lemmy lol. But if you have any ideas, please let me know… I’d love to test it out if I have a good use case.

      • BlackLaZoR@kbin.run
        link
        fedilink
        arrow-up
        3
        arrow-down
        1
        ·
        1 month ago

        But if you have any ideas

        Both my avatar and channel cover are made with AI models - so this is a good start.

        IMO the biggest potential is indie game dev - AI image generation is amazing for static backgrounds, character design, and with certain loras it absolutely shreds pixelart - I even saw entire workflows for building pixelart animations (I think it was for ComfyUi tho).

        Also local image models are uncensored so… porn XD

      • Zos_Kia@lemmynsfw.com
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 month ago

        If you like to write, I find that story boarding with stable diffusion is definitely an improvement. The quality of the images is what it is, but they can help you map out scenes and locations, and spot visual details and cues to include in your writing.

  • CallMeButtLove@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 month ago

    Is there a way to host an LLM in a docker container on my home server but still leverage the GPU on my main PC?

    • azl@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 month ago

      You would need to run the LLM on the system that has the GPU (your main PC). The front-end (typically a WebUI) could run in a docker container and make API calls to your LLM system. Unfortunately that requires the model to always be loaded in the VRAM on your main PC, severely reducing what you can do with that computer, GPU-wise.

  • chip@feddit.rocks
    link
    fedilink
    English
    arrow-up
    3
    ·
    1 month ago

    I also recently got into selfhosting LLM. Having an AMD card meant I had to scourge for solutions since everything expects to have CUDA suppport which means having Nvidia cards. Koboldcpp has a fork with ROCM support which works on my machine, so I’m content with that for now.

    • Toribor@corndog.social
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 month ago

      Do you have any links or guides that you found helpful? A friend wanted to try this out but basically gave up when he realized he’d need an Nvidia GPU.

  • chasingtheflow@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 month ago

    Very cool! You can use something like Tailscale to access your local services remotely without exposing them to the internet.

    • chagall@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 month ago

      Open WebUI now has a docker environment variable so you can, by default, turn off the login page. You just declare it when you’re spinning up the container and you’re good to go.

  • Decronym@lemmy.decronym.xyzB
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    1
    ·
    edit-2
    1 month ago

    Acronyms, initialisms, abbreviations, contractions, and other phrases which expand to something larger, that I’ve seen in this thread:

    Fewer Letters More Letters
    NVR Network Video Recorder (generally for CCTV)
    PSU Power Supply Unit
    VPN Virtual Private Network

    3 acronyms in this thread; the most compressed thread commented on today has 12 acronyms.

    [Thread #917 for this sub, first seen 12th Aug 2024, 07:15] [FAQ] [Full list] [Contact] [Source code]