show code js

2024年3月2日 星期六

Use Ollama with Chatbot-Ollama to use local LLM files.

Just go

  • Download and Install ollama from ollama.com or ollama.ai
    • Open http://127.0.0.1:11434/ to show [ollama is running]
    • Use PowerShell to use ollama
      command: ollama list , to list all model
      command: ollama rm modelname , to del model
  • Install docker manager and use it to install chatbox-ollama
    • Open http://127.0.0.1:3000/
  • Use already download model files
    • Create a Modelfile as your_modelname, and config LLM path like below
      FROM c:\path\your_modelname.gguf
      SYSTEM ""
    • Open powershell and goto your_modelname file of directory, run command like below
      ollama create your_modelname -f ./your_modelname
      ollama run your_modelname(>> /bye to exit)
  • Open http://127.0.0.1:3000/zh , chat with llm

沒有留言:

張貼留言