Obsidian with Ollama

Поділитися
Вставка
  • Опубліковано 23 гру 2024

КОМЕНТАРІ • 31

  • @mrashco
    @mrashco 6 місяців тому +3

    Awesome! I've been using Backyard AI for local LLMs. Obsidian is new to me (switched from Notion) and Ollama looks PERFECT for integrating notes and AI. Thanks for the great video!

    • @the-writer-dev
      @the-writer-dev  4 місяці тому

      Thanks for the support! As a developer and solopreneur, AI and Obsidian are my essential tools so I will keep uploading about them!

  • @JonDAlessandro
    @JonDAlessandro Місяць тому

    Why isn't there an ollama (local) option in my default models?

  • @erinray878
    @erinray878 5 місяців тому +1

    Thank you very much for this video! I just downloaded Obsidian a couple days ago, and was looking for free Copilot alternatives. Do you have any recommendations for the Whisper plugin? (Alternatives, or ways to us a local LLM like in this tutorial)? Thanks again!

  • @SimonCliffordconnect
    @SimonCliffordconnect Місяць тому

    I cant get this to work on a windwos machine. Please create an updated video, using Llama, with Docker and setup on a windows machine

  • @peterbizik224
    @peterbizik224 6 місяців тому +1

    Thank you for the video, this is long time on my todo list, but rather have it on homelab server instead of locally, (not sure if possible).
    Please, what are optimal hw requirements, more cpu, or more memory? What was the bottleneck, It was a bit slow with response locally? Any reason?

    • @the-writer-dev
      @the-writer-dev  6 місяців тому

      Homelab server sounds really cool! Like I said, I'm using 2020 Mac Air m1 so I experienced slow performance when I used a bigger model like llama3. Phi was working great though

  • @envoy9b9
    @envoy9b9 3 місяці тому

    how do you get it to read pdf's?

  • @radonryder
    @radonryder 7 місяців тому +1

    Excellent video!
    Going to try this out.

    • @the-writer-dev
      @the-writer-dev  7 місяців тому

      Thanks and let me know your experience!

  • @daedalusjones4228
    @daedalusjones4228 3 місяці тому

    Great video. Thank you, brother!
    I, too, installed Llama 3 on my machine, and the program/machine was so slow, it just seemed to freeze. It would EVENTUALLY eke out a response, but...no. So thanks for the intel about Phi3, especially!

  • @HiltonT69
    @HiltonT69 7 місяців тому +1

    What would be awesome is for this to be able to use an Ollama instance running in a container on another machine - that way I can use my container host for Ollama with all it's grunt, and keep the load off my smaller laptop.

    • @the-writer-dev
      @the-writer-dev  7 місяців тому +1

      That is an interesting idea..! Thanks for the feedback I will look into this to see it’s possible

    • @tomw0w
      @tomw0w 6 місяців тому +1

      ​@@the-writer-dev i have been experimenting with running Ollama on a Docker container using Proxmox LXC. After configuring the Ollama base URL field with my server's URL on Obsidian copilot, everything works like a charm

  • @siliconhawk
    @siliconhawk 7 місяців тому

    what is the hardware requirements to run models locally.

    • @TheGoodMorty
      @TheGoodMorty 7 місяців тому

      It can run CPU-only, it can even run on a Raspberry Pi, it's just going to be slow if you don't have a beefy GPU. Pick a smaller model and it should be alright. But unless you care about being able to customize the model in a few ways or having extra privacy with your chats, it'd probably just be easier to use an external LLM provider

    • @coconut_bliss5539
      @coconut_bliss5539 7 місяців тому +2

      I'm running Llama3 8B model with Ollama on a basic M1 Mac with 16gb RAM - it's snappy. There is no strict cutoff for hardware requirements - if you want to run larger models with less RAM, Ollama can download quantized models which enable this (for a performance tradeoff). If you're on PC with GPU, you need 16GB of VRAM to run Llama3 8B natively. Otherwise you'll need to use a quantized model.

  • @nevilleattkins586
    @nevilleattkins586 6 місяців тому +2

    If you get an error when you try to run the serve command about port already being use then run 'osascript -e 'tell app "Ollama" to quit''

    • @reddezimen
      @reddezimen 4 місяці тому

      says "osascript: command not found"

  • @elgodric
    @elgodric 7 місяців тому

    Can this work with LM Studio?

    • @the-writer-dev
      @the-writer-dev  7 місяців тому +3

      Good question I haven’t played with LM studio. I will and let you know!

  • @etherhealingvibes
    @etherhealingvibes 7 місяців тому

    Copilot needs integration with Groq AI, and Text to speech integration inside chat room.

    • @the-writer-dev
      @the-writer-dev  7 місяців тому +1

      That sounds interesting idea!

    • @etherhealingvibes
      @etherhealingvibes 7 місяців тому

      ​@@the-writer-devI will cover the costs, allowing us to remove WebsUI and solely utilize Ollama or LMstudio for the backend. With LMstudio now featuring CLI command capabilities, it's even more beneficial as it reduces the layers above Copilot. I conducted a test with LMstudio's new feature today, and the Copilot responses were noticeably faster on my low-end laptop. Additionally, we can incorporate groq's fast responses and edge neural voices, which are complimentary.

  • @Ludwig6583
    @Ludwig6583 4 місяці тому

    Thanks for the video. If it says the `address is already in use`, run this exact command: osascript -e 'tell app "Ollama" to quit'

    • @reddezimen
      @reddezimen 4 місяці тому

      says "osascript: command not found"

  • @IFTHENGEO
    @IFTHENGEO 7 місяців тому +1

    Awesome video man! Just sent you connect on LinkedIn

    • @the-writer-dev
      @the-writer-dev  7 місяців тому +1

      Thanks for the support and I will check it out!

  • @VasanthKumar-rh5xr
    @VasanthKumar-rh5xr 7 місяців тому +1

    Good video. I get this message in the terminal while setting the server step 4. >>> OLLAMA_ORIGINS=app://obsidian.md* ollama serve
    The "OLLAMA_ORIGINS" variable in the context provided seems to be a custom
    configuration, and serving files with `ollama` would again follow standard
    Node.js practices:
    1. To set an environment variable similar to "OLLAMA_ORIGINS", you could
    do so within your project's JavaScript file or use shell commands (again
    this is for conceptual purposes):
    I can connect with you through other channels to work on this step.

    • @rjt_y
      @rjt_y 6 місяців тому

      can you please explain more i cant get mine working