shakedown.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
A community for live music fans with roots in the jam scene. Shakedown Social is run by a team of volunteers (led by @clifff and @sethadam1) and funded by donations.

Administered by:

Server stats:

292
active users

#ollama

1 post1 participant0 posts today

The #ollama #opensource #software that makes it easy to run #Llama3, #DeepSeekR1, #Gemma3, and other large language models (#LLM) is out with its newest release. The ollama software makes it easy to leverage the llama.cpp back-end for running a variety of LLMs and enjoying convenient integration with other desktop software.
The new ollama 0.6.2 Release Features Support For #AMD #StrixHalo, a.k.a. #RyzenAI Max+ laptop / SFF desktop SoC.
phoronix.com/news/ollama-0.6.2

www.phoronix.comollama 0.6.2 Released WIth Support For AMD Strix Halo

So, following a blog post[1] from @webology and some guides about #VSCode plugins online, I set up #ollama with a number of models and connected that to the Continue plugin.

My goal: see if local-laptop #llm code assistants are viable.

My results: staggeringly underwhelming, mostly in terms of speed. I tried gemma3, qwen2.5, and deepseek-r1; none of them performed fast enough to be a true help for coding.

So, I did it. I hooked up the #HomeAssistant Voice to my #Ollama instance. As @ianjs suggested, it's much better at recognizing the intent of my requests. As @chris_hayes suggested, I'm using the new #Gemma3 model. It now knows "How's the weather" and "What's the weather" are the same thing, and I get an answer for both. Responses are a little slower than without the LLM, but honestly it's pretty negligible. It's a very little bit slower again if I use local #Piper vs HA's cloud service.

Testing out the newly released #Gemma3 model locally on #ollama. This is one of the more frustration aspects of these LLMs. It must be said that LLMs are fine for what they are, and what they are is a glorified autocomplete. They have their uses (just like autocomplete does), but if you try to use them outside of their strengths your results are going to be less than reliable.

Be careful if you are running an #Ollama web server

According to this article if you run Ollama as a web server, meaning you are running an LLM model locally on your server or home computer, but you have a web portal open to it so people in your organization or home can connect to your server and ask the LLM questions, the Ollama web server is apparently full of security holes. The article mentions three problems:

  • It can leave your computer vulnerable to DDoS attacks from the public Internet
  • The push/pull feature for uploading/downloading models is vulnerable to man-in-the-middle attacks (possibly? as is my understanding)
  • DeepSeek is not a security issue in and of itself, but since DeepSeek is so easy for hobbyists to use, this is causing a larger number of people to use Ollama, increasing the number of people who are vulnerable.

Quoting the article:

the API can be exposed to the public internet; its functions to push, pull, and delete models can put data at risk and unauthenticated users can also bombard models with requests, potentially causing costs for cloud computing resource owners. Existing vulnerabilities within Ollama could also be exploited.

Hackread - Latest Cybersecurity, Tech, Crypto & Hacking News · 7,000 Exposed Ollama APIs Leave DeepSeek AI Models Wide Open to AttackFollow us on Bluesky, Twitter (X) and Facebook at @Hackread
#tech#AI#OpSec