Hello, I have an important question. In general, video cards degrade after long-term work with LLMs, and all because the video cards are not loaded at all at first, and when a request comes from the user, they are instantly loaded to 100%, which is why the video chip degrades. Is it possible to somehow make sure that the video card is constantly at 70-100% load? The simple fact is that the condition of constant 70-100% is more favorable for a video card than from 0% to 100% in an instant. Thanks for your reply (This comment was written with the help of a translator).
This is not true. GPU chips aren't machinery, they don't degrade from load cycling. You are looking for a solution to a problem that doesn't exist. If you're concerned about *thermal cycling* which could in theory wear out your cpu fan and cause a gpu to overheat when the fan fails, then set a power limit on the card to keep temps down.
Brother, can you make a tutorial about LM Studio to do screen sharing with character? With llava or some LLM we can run offline on my rig? I can use llava inside LMStudio, but I can't connect it to sillytavern. I didn't have any problem connect LM studio to Vpet via API.
Can we use LM Studio as backend for vision capable models then connect it to Silly Tavern? I don't understand how to set it in image captioning extension as it's not in the options. Do we need to use an OpenAI-compatible API and set it as Chat Completions in Silly Tavern? Is that possible for LM Studio?
i'm using this too because it's more friendly user interface.
They did just update OpenCl to Vulkan so NICE.
I'm so happy about this update. I have an AMD card unsupported by ROCm but was able to get models running on my GPU thanks to vulkan.
Hello, I have an important question. In general, video cards degrade after long-term work with LLMs, and all because the video cards are not loaded at all at first, and when a request comes from the user, they are instantly loaded to 100%, which is why the video chip degrades. Is it possible to somehow make sure that the video card is constantly at 70-100% load? The simple fact is that the condition of constant 70-100% is more favorable for a video card than from 0% to 100% in an instant. Thanks for your reply (This comment was written with the help of a translator).
This is not true. GPU chips aren't machinery, they don't degrade from load cycling. You are looking for a solution to a problem that doesn't exist. If you're concerned about *thermal cycling* which could in theory wear out your cpu fan and cause a gpu to overheat when the fan fails, then set a power limit on the card to keep temps down.
@@stop_tryharding thanks for information. I appreciate it.
yeah THE GOAT
Brother, can you make a tutorial about LM Studio to do screen sharing with character? With llava or some LLM we can run offline on my rig? I can use llava inside LMStudio, but I can't connect it to sillytavern.
I didn't have any problem connect LM studio to Vpet via API.
Thanks! This looks pretty interesting! Will you cover Jan AI sometime as well?
Wowie look good 👍
Can we use LM Studio as backend for vision capable models then connect it to Silly Tavern? I don't understand how to set it in image captioning extension as it's not in the options.
Do we need to use an OpenAI-compatible API and set it as Chat Completions in Silly Tavern? Is that possible for LM Studio?
ooba with more cool UI. But if one have a potato pc, have a potato pc XD
i use infermatic fo online LLM
For Linux users ollama can be more preferable
Cool, but is it possible to connect ollama and ST?
Please tell me what is the best model for RP that I can run locally on my 3070ti?
ua-cam.com/video/uFcEMfYbh9c/v-deo.html
What are the components of your PC?
But not open source?
Its literally on Github....
@@snintendog
Oh, okay.
@@snintendog no it's not, only the prompt templates are open source. The app is closed.
@@strkn25 The app is just a python wrapper....right cant talk to those that can do basic research.