Just a stranger trying things.

  • 9 Posts
  • 299 Comments
Joined 2 years ago
cake
Cake day: July 16th, 2023

help-circle

  • Same, I rocked a second hand GTx 680 from 2012-2013, which I upgraded to a second hand RTX 3060 12GB for a fantastic price, in 2024. Still rocking a DDR3 platform with the intel i7 4400K. And that’s more than enough for most games with nice graphics on 1680x1050 :) (display probably 15 years old too). Eventually, I will be looking for some other second hand components to upgrade the rest of the system, but it does everything more than well enough.








  • I don’t understand in what circumstances anyone would like to use link shorteners? I can only find reasons why not to use them:

    • subject visitors to surveillance
    • destination under the control of a third party (potential for ransom for the author, like we see here, and potential of ads for visitors like we see here)
    • obfuscation of the actual destination
    • how long will the redirect be valid for? It could be deemed unviable for the company to continue support of the redirect, thus rendering the destination inaccessible from all places where the shortened link is used.
    • more…?



  • One thing which I find useful is to be able to turn installation/setup instructions into ansible roles and tasks. If you’re unfamiliar, ansible is a tool for automated configuration for large scale server infrastructures. In my case I only manage two servers but it is useful to parse instructions and convert them to ansible, helping me learn and understand ansible at the same time.

    Here is an example of instructions which I find interesting: how to setup docker for alpine Linux: https://wiki.alpinelinux.org/wiki/Docker

    Results are actually quite good even for smaller 14B self-hosted models like the distilled versions of DeepSeek, though I’m sure there are other usable models too.

    To assist you in programming (both to execute and learn) I find it helpful too.

    I would not rely on it for factual information, but usually it does a decent job at pointing in the right direction. Another use i have is helpint with spell-checking in a foreign language.


  • Regarding photos, and videos specifically:

    I know you said you are starting with selfhosting so your question was focusing on that, but I would like to also share my experience with ente which has been working beautifully for my family, partner and myself. They are truly end to end encrypted, with the source code available on github.

    They have reasonable prices. If you feel adventurous you can actually also host it yourself. They have advanced search features and face recognition which all run on device (since they can’t access your data) and it works very well. They have great sharing and collaborating features and don’t lock features behind accounts so you can actually gather memories from people on your quota by just sharing a link. You can also have a shared family plan.






  • To run the full 671B sized model (404GB in size), you would need more than 404GB of combined GPU memory and standard memory (and that’s only to run it, you would most probably want it all to be GPU memory to make it run fast).

    With 24GB of GPU memory, the largest model which would fit from the R1 series would be the 32b-qwen-distill-q4_K_M (20GB in size) available at ollama (and possibly elsewhere).



  • The Hobbyist@lemmy.ziptoSelfhosted@lemmy.worldI installed Ollama. Now what?
    link
    fedilink
    English
    arrow-up
    27
    arrow-down
    1
    ·
    edit-2
    2 months ago

    Ollama is very useful but also rather barebones. I recommend installing Open-Webui to manage models and conversations. It will also be useful if you want to tweak more advanced settings like system prompts, seed, temperature and others.

    You can install open-webui using docker or just pip, which is enough if you only care about serving yourself.

    Edit: open-webui also renders markdown, which makes formatting and reading much more appealing and useful.

    Edit2: you can also plug ollama into continue.dev, an extension to vscode which brings the LLM capabilities to your IDE.