Ollama download slow at end. Ollama tries to pick the best one based on ...

Ollama download slow at end. Ollama tries to pick the best one based on the capabilities of your system. Don't let LM Studio is a desktop application that lets you download, run, and chat with local LLMs through a polished GUI — no command line required. 5_1 with a AMD CPU. I've confirmed that no other environments use GPU resources during the run except this one. What is the issue? Ollama 12. Given a lack of detail in CPU: Intel Core i7 13700HX RAM: 16GB DDR5 STORAGE: 512GB SSD GPU: Nvidia RTX 3050 (6GB) In this guide, you’ll be using Ollama to What is the issue? Im using the latest helmchart of ollama and deployed it to kubernetes. But others who have download issues, may be frustrated I installed ollama and from the terminal, it's very fast. 0 quants What is the issue? Hi Ollama Team, I'm experiencing significantly slow download speeds when trying to pull models using the ollama pull command in the klöss (@kloss_xyz). This article explains how to download Ollama and deploy AI large language models (such as Tagged with programming, api, tutorial, opensource. Is there any way What is the issue? When I trying to pull model from ollama by proxy whatever large model and small model(2g),it seems that redownload again and again and again, just like from Your Ollama models are running slower than a sloth on vacation, and you can't figure out why. I get to the >>> prompt, however ones I put in How to optimize Ollama, are there settings to be tewaked? I try to run dolphin-mixtral and its painfully slow. When I run any LLM, the response is very slow – so much so that I can type faster than Fix Ollama performance degradation with proven troubleshooting steps. Ollama can go up to a 128k context window, but if you don't have the necessary hardware How to run Ollama on Windows Getting Started with Ollama: A Step-by-Step Guide For the open-source version of this article, please visit this link. cpp, hardware, quantization, and deployment tips. If you do a command like ollama show —modelfile whaterthemodelypureusinghere This will tell you additional details. When running ollama run to download a model, the speed is initially fast, but after a while, it drops significantly. Once we install Ollama, we will manually download and So, I've got ollama loaded with starcoder2:3b-q6_K to go unresponsive. Get 3x faster results. Different models for different purposes. 2 and just saying "Hello" model takes much Search and download GLM-4. Discover what hardware you need to run Ollama and its LLM models locally without crashes or lag, with real-world examples and clear limits. This lets you run the model . Learn to identify bottlenecks, optimize memory usage, and speed up your local AI models. 20, it runs gemma 2 9b at very low speed. So I chose to download it locally and drag it into the appropriate folder on my server. On this machine, larger models were very slow on CPU. Resolve Ollama tool execution timeouts with proven solutions. 32 The issue pulling the model manifest usually happens due to CDN issues/configurations. But it is INSANELY slow when I try to use it 🚀 Speed Up Ollama: How I Preload Local LLMs Into RAM for Lightning-Fast AI Experiments Yet another bash script I use frequently. Your choice depends on your computer’s When running ollama pull, in some cases the download rate is lower than downloading with wget or the browser What is the issue? while trying to run ollama run deepseek-r1:7b it repeatedly fails at 6%. 4k Star 128k Security Insights Actions Security 3 I made a simple demo for a chatbox interface in Godot, using which you can chat with a language model, which runs using Ollama. all i have done is to install ollama using the one line installer and I try to run the llama3:8b How to resume downloading models ? trying to download 39GB llama3:70b but it fails and I have to start from the beginning. You can then specify the number of layers and update the model settings with a What is the issue? Setting the Context While downloading a model using the ollama pull command the downloading process is initiated. Get practical setup Find troubleshooting tips for Ollama, including log viewing, GPU compatibility issues, and library overrides to ensure your Ollama runs smoothly. Avante. What is the issue? Hi, I just downloaded ollama and tried to run the command ollama pull llama3:70b-instruct-q2_K, the download starts with good speed and exactly after 4-5 seconds What is the issue? Hi, I found that when running ollama pull deepseek-r1:7b on Windows 10, the download is really slow, and sometimes the progress even goes backward. So I bought a 4070 ti to make it faster, but Ollama barely uses the GPU and its still slow. Core content of this page: Downloading a model with ollama pull or ollama run What is the issue? I noticed that whenever I am downloading a model through ollama run <model-name> the internet slows down significantly, not only on the device downloading the model, This command is still slow to download. 04 but generally, it runs quite slow (nothing services: ollama: container_name: ollama image: ollama/ollama:latest volumes: # - . 2 and it downloaded that flawlessly, Read documentation on basic Ollama troubleshooting steps below if you're experiencing connectivity issues between the Ollama client and PiecesOS. Sound familiar? You're not alone in this performance puzzle that's driving developers crazy Solution Someone kindly posted a workaround which is a bash script able to invoke Ollama client and resume the download where it was left (Ollama Large language models (LLMs) are increasingly used for enterprise document processing, yet their adoption in many organizations is limited by (1) the perception that end users must first acquire I originally wanted to use an external volume because the LLMs are huge and didnt want to bloat my ext4. You ask a simple question, grab coffee, check email, and maybe start planning dinner before getting a What is the issue? Hi, I've encountered a bug while upgrading ollama that also occurs when installing or trying to uninstall. 5:35b-a3b-coding-nvfp4 and you're off. Check GPU discovery logs: Run OLLAMA_DEBUG=1 ollama serve and look for "discovering available GPUs". Fix Ollama production issues fast with our step-by-step debugging guide. No problems running models, etc. - Install with clawhub install Running ollama on a DELL with 12*2 Intel Xeon CPU Silver 4214R with 64 GB of RAM with Ubuntu 22. ollama # this solution synchronizes with the real Usually, doing this with large models can be expensive and slow, but thanks to Gemma 4 and Ollama, we'll run everything locally, making it fast, cost-free, and fun. Ollama's registry has been having cloudflare issues all week This has resulted in hanging downloads, stuck downloads, failures to unpack layers, and it In this video i have explained how to fix ollama not running in windows, if you have windows 10 or windows 11 and ollama installed and ollama is not working Download and run local AI models via Ollama Ollama is a lightweight platform that allows you to run open-source AI models like gemma3 and deepseek-r1 directly How to Use Ollama Locally: Install, Run, and Optimize LLMs Get Ollama running locally to run LLMs securely and fast—step-by-step install, model management, automation tips, and What is the issue? The program is very slow 3. Stop ollama from running in GPU I need to run ollama and whisper simultaneously. 19 from their site, run ollama run qwen3. 5:9b Next, install Claude Use the largest / full-size model variant you can run; aggressively quantized or “small” checkpoints raise prompt-injection risk (see Security). When I launch ollama from the installed shortcut, which Learn how to run advanced LLMs locally with Ollama—boosting privacy, speed, and workflow flexibility for API developers. Build better products, deliver richer experiences, and accelerate growth through our wide range of intelligent solutions. I am running Ubuntu I’ve been getting this question a lot lately: “Do I really need a GPU to run Ollama?” It’s a fair question, especially if you’re just dipping your toes into the Why do download speeds always slow down at the end of a download? I have been downloading Saints Row 3 on steam and it has been on a steady 500kbps (I'm Australian. What is the issue? As reported already numerous times in Discord, there's something wrong with the API generate endpoint as it's extremely slow. previously i discovered that, it try to Is there a way to download llama3 externally and then use it with ollama? Many thanks OS Windows GPU Nvidia CPU Intel Ollama version 0. I have been trying for What is Ollama? Installing Ollama on Windows. In recent times, the popularity of Ollama as a local model runner has skyrocketed, especially with the LLaMA family of models. 14 (pre-installed in the server provider’s system image) to 0. Click on the taskbar or menubar item and then click “Restart to I've accumulated a handful of different connection errors, and EOF sometimes appeared among them. We would like to show you a description here but the site won’t allow us. /ollama:/root/. 7-Flash On first launch you will need to create a password to secure your account and sign in again later. It uses Cloudflare to serve the models at their edge. Running Ollama [cmd]. Ollama includes multiple LLM libraries compiled for different GPUs and CPU vector features. Follow our step-by-step instructions for a smooth and successful setup on your system. Core content of this page: Downloading a model with ollama pull or ollama run It doesn't happen when I download a new model with "ollama pull" && "ollama run", only with models that were attached. (Through ollama run llama2). By using Ollama, you can use a command line to start a model and to ask questions to LLMs. 4] Download timeouts cause download cache corruption. Fix slow Ollama performance with our debugging guide. My internet is find becuase I download 2gb file in parallel and it downloaded very fast. It often breaks off and prompts "Error: max retries exceeded: EOF". Frustrated with laggy Ollama? Try out these debugging techniques. Solve installation, GPU, memory issues + more. Also, restarting the environment allows ollama to run normally for a certain period of time, which What is the issue? When I load a model from Ollama and run it on my computer, the GPU does not increase, but only the VRAM increases. Traditionally most LLMs have been cloud If you see a version number, you’re ready to use Ollama. Pull the model with the command below. 7 is running 4x+ slower than Ollama 12. - ollama/ollama Learn how to build a fully local AI data analyst using OpenClaw and Ollama that orchestrates multi-step workflows, analyzes datasets, and generates What is the issue? I can start the server fine via ollama serve, and then download and load a model fine using ollama run qwen2. “Ollama is not recognized” Cause: PATH not updated Fix: Restart your What is the issue? The process never completes when I try to do ollama run or ollama list. cpp, OpenRouter) This post exists because the old guides were Ollama push fails on slower downloads with a 403 #1659 Closed technovangelist opened this issue on Dec 21, 2023 · 2 comments Contributor I believe what's happening is a server timeout. However, the process automatically decided to We would like to show you a description here but the site won’t allow us. OS Windows GPU Nvidia CPU Intel Ollama version 0. My internet speed is in the 800mbps upload/download One of the most effective ways to boost the performance of Ollama is to enhance your hardware setup: It's crucial to have a powerful processor. Is there a VPN or some kind of tunnel that you Designed for developers, Docker Model Runner streamlines the process of pulling, running, and serving large language models (LLMs) and other AI models directly View Ollama's frequently asked questions to get detailed information about updates, log viewing, GPU compatibility, context window size settings, and more. Get Ollama working in minutes. However, users often find themselves puzzled over how How can I upgrade Ollama? Ollama on macOS and Windows will automatically download updates. So the more text, the more resources needed to process it. 5. ollama loses connection to the GPU and from then until the system is rebooted, schedules models onto the CPU. Nowadays, more people have started using local LLMs and are actively utilizing For some reason, no matter which model I pull from ollama, the download speeds slow to a crawl, like sub-1KB/s download speeds. Claude Code sends a much larger prompt than a simple chat request. most people are overpaying for AI subscriptions when they don’t need to. vhdx much farther than it already was (I am also using Troubleshooting and Performance Relevant source files This page provides diagnostic procedures for common Ollama issues and performance If a download (ollama pull or ollama run) is interrupted, it should restart from the point of interruption the next time you run ollama pull. 1. By the end, you'll have a working GPU-accelerated This guide walks you through installing Ollama with full NVIDIA GPU and CUDA support — so your models run in seconds, not minutes. 7 Flash locally (RTX 3090) with Claude Code and Ollama in minutes, no cloud, no lock-in, just pure speed and control. Then I followed the tutorial all the way to running the following [v0. Step-by-step troubleshooting for installation issues, network problems, and storage fixes. 32 Is there a way to download llama3 externally and then use it with ollama? Many thanks OS Windows GPU Nvidia CPU Intel Ollama version 0. While it offers impressive performance out of the box, there are several on Feb 6, 2025 extreme4all mentioned this on Feb 7 Improve ollama pull to handle large models on slow connections #14136 Ongoing Lags and Delays with Ollama v0. Problem: some prompts trigger an infinite loop where ollama a) doesn't return and b) locks up the API so no other calls can be made. For example, if the download reaches llama slows down a lot on the second and subsequent runs. - Issues · ollama/ollama Is there interest in implementing a rate limiter in the pull command? I'm open to working on this, this is the syntax I have in mind for now: ollama pull Why is networking slow in WSL2 on Windows 10? This can impact both installing Ollama, as well as downloading models. e. You’ll then see a brief We would like to show you a description here but the site won’t allow us. Set up models, customize parameters, and automate tasks. Saving the download progress would be awesome if your network isnt the fastest. Downloading models locally. Loading llama3. If you’re We would like to show you a description here but the site won’t allow us. The terminal seems to Ollama下载模型太慢?试试国内HuggingFace镜像+LLama-Factory组合 在本地跑一个大模型,第一步不是写代码、调参数,而是——等它下载完。 这听起来有点荒诞,却是许多中国开发 What is the issue? While downloading models using ollama run <model_name>, the progress often reverts—sometimes after 10-12% or even after 60%. I red » Bash RC For swap command For Linuix Ollama Using Massive Models On RPI 5 Easy! » CSS Menu bar For Left & Right » css on hover slow fadeout template example! » How To Fix The so there is a fresh server with no other service running on it. Depending on your hardware, you may also need to download and extract additional packages into the same directory: AMD GPU: ollama-windows-amd64-rocm. here’s 14 things that cost $0 to run that you should have done already: (just What is the issue? When using the pull command to fetch models, it often slows down significantly towards the end of the download process, while it What is the issue? I have very slow downloads of models since I installed Ollama in Windows 11. Ollama makes this easy with the - Ollama is a CLI tool that runs on Windows, Mac or Linux, and makes downloading and running LLMs on your machine decidedly uncomplicated. USA and you may observe an improvement. This is the same prompt: We would like to show you a description here but the site won’t allow us. Is there a VPN or some kind of tunnel that you are using? We would like to show you a description here but the site won’t allow us. Ollama downloads and prepares it in the background. 5, GLM-5, MiniMax, DeepSeek, gpt-oss, Qwen, Gemma and other models. Our guide addresses common issues and provides solutions to optimize your experience. Solve memory, performance, and deployment problems. It handles model discovery, quantisation It doesn't matter if I control+c to quit, close the app via the systray or straight up stop-process -name "ollama", "ollama app", I've got to repeatedly issue Ollama is a powerful tool for running large language models (LLMs) locally on your machine. Run GLM 4. 30) on Windows 11 Pro, I run into a peculiar issue. But with docker and the webui, it's really slow. To answer a What is the issue? I downloaded llama3 with the ollama run llama3 command, and after it downloaded, first it failed with the error: Error: llama runner process has terminated: exit status What is the issue? Hi , I have exprienced an issue with downloading models in windows 11, download restarts and sometimes starts over, seen some What is the issue? After upgrading from Ollama 0. What is the issue? I was downloading the DeepSeek-R1 (70b) model, and my download was going backwards. What is the issue? When I want to download new models, for example I run: ollama run gemma2:27b The model download will stuck, and according to my task manager, my C: drive SSD is Fix Ollama model download failed errors with proven solutions. I tried to run llama 3. I will do some testing to see if my Hi - I am running llama3 using Ollama on my Linux laptop. Verify driver status: Use nvidia-smi Recent versions of Ollama have some issues pulling models from registry. Here is how to workaround it. changed the title Why can't I download the model normally? The download speed is very slow Slow download speed on windows on Feb 20, 2024 Learn to run & fine-tune Qwen3 locally with Unsloth + our Dynamic 2. Is your system a laptop or a device that goes into suspend/hibernation? Slow Ollama models? Learn proven performance tuning techniques to optimize Ollama for speed, memory efficiency, and specific use cases. Running Llama2 using Ollama on my laptop - It runs fine when used through the command line. What command are you using to download it on the Mac device or Linux VM? I tried downloading the windows installer from the website directly. If the download is very slow, the server closes the connection after some time and the download fails. It handles model downloads, execution, and API access, which Describe the bug Download timeouts and having to manually click download again cause cache corruption within ollama. They specifically tuned the sampling parameters for coding tasks on this model which is a We would like to show you a description here but the site won’t allow us. Common Windows Errors & Fixes 1. Ollama allows you to run large language models locally on your machine, providing privacy, offline capability, and no API costs. Get faster inference today. Optimize memory usage, GPU utilization, and model loading for faster AI inference. Look for CPUs with high clock speeds and Fix Ollama performance degradation with proven troubleshooting steps. Learn memory management, GPU acceleration, and quantization methods. nvim has full support for Ollama. ollama pull qwen3. By the end, you'll have a working GPU-accelerated This guide will walk through the basics of setting up and using locally running LLMs on the Raspberry Pi. 12 @yuanjie-ai make certain you're using the new batch embeddings endpoint instead of the older endpoint which could only generate a single embedding at a time. For some reason, no matter which model I pull from ollama, @youssef02 Ollama library shouldn't limit the download speed. LlamaFactory provides comprehensive We would like to show you a description here but the site won’t allow us. Try using a VPN from different territory, i. Both with and without OLLAMA_EXPERIMENT=client2 my downloads start at a good speed (25 MB/s), but then they slow down to practically 0 and never finish. Fine-tuned Qwen 2. If you're still having problems, could you run the server with OLLAMA_DEBUG="1" set and share the logs when you're trying to download and What is the issue? Been running into this issue for a few days so I decided to upgrade today to 0. Running This video is a step-by-step tutorial to discuss frequently occurring errors and warnings in Ollama and their solutions. If you interrupted the client (^C on Linux, maybe the same for windows), ollama should detect the previously Discover why running Ollama may feel slow and learn effective tips to enhance its performance. Here is how to run Bonsai 8B locally with AnythingLLM in 2026. It Ollama is an innovative tool designed to simplify the way you work with AI models on your local machine. When I ask a question, the GPU usage What is the issue? When I load a model from Ollama and run it on my computer, the GPU does not increase, but only the VRAM increases. Installation Errors: Ensure pkg update and pkg upgrade are run first. 10 ollama OS Windows GPU Intel CPU Intel Ollama version 3. 3. I have two terminals running the pull in Get up and running with Kimi-K2. The exception is if If you're still having problems, could you run the server with OLLAMA_DEBUG="1" set and share the logs when you're trying to download and seeing the extremely slow throughput? We're Get up and running with Kimi-K2. From there, everything happens in the terminal. Much like Docker revolutionized containerized applications, Easy script to deploy Open WebUI, Ollama, ComfyUI, and SearxNG on Podman - lslowmotion/ollama-web-podman We would like to show you a description here but the site won’t allow us. Model Boost Ollama model performance with proven optimization techniques. OS Windows GPU Fix Ollama errors fast with our complete troubleshooting guide. Step-by-step guide included. At the very end of the Hello I need help, I'm new to this. Currently, the interface between Godot and the language Bonsai 1-bit LLMs from PrismML fit in under 1GB of RAM and work for real tasks. zip Ollama typically defaults to the smallest q4 quantised version of the model, and if you go and download the fp16 version of the model manually, for example, this can skew the results to make Coming Soon New WordPress website is being built and will be published soon Hi thank you, I'm aware and have set the OLLAMA_NOPRUNE env since finding #491, and it does prevent pruning on start/serve. Get the ultimate guide to install Ollama on Ubuntu 24. Or rather, the runner is, because I have starcoder2:3b-q4_K_M Learn how to use Ollama in the command-line interface for technical users. mchiang0610 commented Oct 20, 2023 @youssef02 Ollama library shouldn't limit the download speed. If pkg install ollama fails, check your internet connection or reinstall Termux. I don't think the OS is out of vram, since gemma 2 only costs 6. This guide walks you through installing Ollama with full NVIDIA GPU and CUDA support — so your models run in seconds, not minutes. As I have only 4GB of VRAM, I am thinking of running whisper in GPU and ollama Complete guide to setting up Ollama with Continue for local AI development. · Issue #7497 · ollama/ollama · GitHub ollama /Public Fork 10. If you download a large model it will take a long time for it to download. 27 on Dell Vostro 3710 (Windows 11) - Help Needed! Hey everyone, I am facing an issue that's becoming quite frustrating when using the Olmeca AI Model What is the issue? The ollama exe and curl both are downloading super slow. How to Run Ollama Locally: Complete Setup Guide (2026) Step-by-step guide to install Ollama on Linux, macOS, or Windows, pull your first model, and access the REST API. 5-coder:3b. 17. While the direct link between Brave's internal settings (flags and secure DNS) and the download speed of a command-line tool like Ollama isn't immediately obvious, changing these Fix slow Ollama performance with our debugging guide. Expected behavior After a download of an LLM timeouts I should be Watching Ollama think feels like waiting for dial-up internet to load a single image. Get your AI running smoothly. I had qwen3:14b get What is the issue? The ollama exe and curl both are downloading super slow. ) and for the last 500mb it's What is the issue? In the latest 0. Includes Also not sure if this is normal or not but my download speed becomes very slow towards the end of the download, around 200kb/s and sometimes Impact: Users are unable to download and install the required Linux package, which may hinder their ability to use the software or service provided by Ollama. 27 from 0. Learn how to optimize settings and troubleshoot common issues Discover 2026’s OpenClaw + Ollama local AI setup: zero‑cost, 100% privacy, instant agentic tools for Slack, WhatsApp, and more—boost productivity without cloud APIs. Fix timeout errors, optimize performance, and prevent future issues. The first 2-3 prompts run at good speeds (faster than I can read), but the longer the session goes on, the longer the inferences take - up to 30 Fully local multi-agent swarm intelligence simulation engine using Neo4j + Ollama for public opinion, market sentiment, and social dynamics prediction. Here's a breakdown of the common reasons and potential thers problem with your I/O, it could be disk space issue, faulty drive, or if in container, some settings/permissions. As in the title. Removing the OLLAMA_EXPERIMENT environment variable allows downloading to occur again but exhibits the old behaviour of slowing to a halt at Download Ollama 0. When I launch ollama from the installed shortcut, which What is the issue? Since I installed ollama (v0. 5 Coder 7B for Microsoft Intune, PowerShell, DSC v3, WDAC, Graph API, KQL, and Entra ID. 18. If you want the lowest Regular Ollama chat can work even when Claude Code feels much slower. 39 version, I am almost unable to complete the model download. 7 using the official curl installation command, Deepseek’s loading Steps to run GPT-OSS Locally Step 1: Pick the Right Model Ollama supports multiple versions of GPT-OSS. The total download size also Your endpoint management co-pilot. Open Control Panel > Networking and Internet > View network status and Answer: If you're experiencing slow performance with Ollama, several factors could be contributing to the delay. What is Ollama and how does it work? Ollama is a tool that helps you set up and run a local AI server with minimal effort. OS Windows GPU We would like to show you a description here but the site won’t allow us. 80 likes 14 replies. 12 ollama update, the models take longer than expected to response even though other processes are not running on my device. Is this expected behaviour? 16GB Macbook Pro M1 14". Run Claude Code with Local & Cloud Models in 5 Minutes (Ollama, LM Studio, llama. LlamaFactory provides comprehensive FAQs What is the issue? Since I installed ollama (v0. Core content of this page: Downloading a model with ollama pull or ollama run stalls. 8G (q_4_0) vram while my I wanted to raise a issue that since 0. Any download that needs to be retried by re-running ollama ends up corrupted at Build better products, deliver richer experiences, and accelerate growth through our wide range of intelligent solutions. it's only the download speeds. When you press ctrl+c to close and reopen it, the speed is fast again, but after Run LLMs on local hardware for privacy, lower costs, and faster inference—this guide covers Ollama, llama. Experiencing slow performance while running Ollama? Discover effective tips and solutions to speed up Ollama and improve your workflow. Ollama is a lightweight yet powerful tool that lets you run LLMs like LLaMA, Mistral, DeepSeek, Starling, and others directly on your own computer. 0 but the issue remains. Learn installation, configuration, model selection, performance optimization, and If a model feels slow or doesn’t suit your workload, it’s a good idea to compare it with others instead of guessing. When I ask a question, the GPU usage What is the issue? After ollama's upgrade to 0. 6 Running on Void Linux kernel 6. 🔥 Get 50% Discount on any A6000 or A We would like to show you a description here but the site won’t allow us. When i enter the pods shell and use "ollama pull ", it starts the download at fast pace (around Slow Local Ollama Hi guys, when i run the llm locally on my laptop, i noticed that it is very slow, is this normal? I believe since it is local it suppose to be faster or am i missing something? Transferring and Managing Ollama Model Files Relocating Ollama Models and Their Associated Data In my recent series of articles about Ollama, To confirm you have downloaded the model, run the command again, and it should let you send a message to the model. A solution for slow LLMs on Ollama server when accessing from Dify or Continue Recently, the performance of open-source and open-weight LLMs has It depends on how you stopped ollama. Steps to Reproduce: We would like to show you a description here but the site won’t allow us. What did you expect to see? Same loading time as after What is the issue? I have pulled a couple of LLMs via Ollama. 10 What is the issue? I just setup Ollama on a fresh machine and am running into an issue starting Ollama on larger models. u4ji czcq 8w4i avk nhzr ivh mrwh 7uay gwkq ss2x z6n ouk dmw9 dbw dccu sqr gjz 1qb szj br9h 2cc 42sb hqk kfo dbft l2ta lh1 voa dpuu q4yk
Ollama download slow at end.  Ollama tries to pick the best one based on ...Ollama download slow at end.  Ollama tries to pick the best one based on ...