About 50 results
Open links in new tab
  1. ollama - Reddit

    r/ollama How good is Ollama on Windows? I have a 4070Ti 16GB card, Ryzen 5 5600X, 32GB RAM. I want to run Stable Diffusion (already installed and working), Ollama with some 7B models, maybe a …

  2. How to make Ollama faster with an integrated GPU? : r/ollama - Reddit

    Mar 8, 2024 · How to make Ollama faster with an integrated GPU? I decided to try out ollama after watching a youtube video. The ability to run LLMs locally and which could give output faster amused …

  3. Request for Stop command for Ollama Server : r/ollama - Reddit

    Feb 15, 2024 · Ok so ollama doesn't Have a stop or exit command. We have to manually kill the process. And this is not very useful especially because the server respawns immediately. So there …

  4. Ollama GPU Support : r/ollama - Reddit

    I've just installed Ollama in my system and chatted with it a little. Unfortunately, the response time is very slow even for lightweight models like…

  5. Local Ollama Text to Speech? : r/robotics - Reddit

    Apr 8, 2024 · Yes, I was able to run it on a RPi. Ollama works great. Mistral, and some of the smaller models work. Llava takes a bit of time, but works. For text to speech, you’ll have to run an API from …

  6. Ollama is making entry into the LLM world so simple that even ... - Reddit

    I took time to write this post to thank ollama.ai for making entry into the world of LLMs this simple for non techies like me. Edit: A lot of kind users have pointed out that it is unsafe to execute the bash file to …

  7. Training a model with my own data : r/LocalLLaMA - Reddit

    Dec 20, 2023 · I'm using ollama to run my models. I want to use the mistral model, but create a lora to act as an assistant that primarily references data I've supplied during training. This data will include …

  8. How to Uninstall models? : r/ollama - Reddit

    Jan 10, 2024 · That's really the worst. To get rid of the model I needed on install Ollama again and then run "ollama rm llama2". It should be transparent where it installs - so I can remove it later. Meh..

  9. Best Model to locally run in a low end GPU with 4 GB RAM right now

    I am a total newbie to LLM space. As the title says, I am trying to get a decent model for coding/fine tuning in a lowly Nvidia 1650 card. I am excited about Phi-2 but some of the posts here indicate it is …

  10. Multiple GPU's supported? : r/ollama - Reddit

    Mar 15, 2024 · Multiple GPU's supported? I’m running Ollama on an ubuntu server with an AMD Threadripper CPU and a single GeForce 4070. I have 2 more PCI slots and was wondering if there …