Activity
Mon
Wed
Fri
Sun
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
Feb
Mar
Apr
What is this?
Less
More

Memberships

Shipping Skool

202 members • $99/month

Digital Degens

393 members • $525

AI Money Lab

70.9k members • Free

Tech Snack University

18.6k members • Free

Zero2Launch AI Automation

5.5k members • $1/m

5 contributions to Shipping Skool
Run 397B model on 48GB RAM
First Google’s TurboQuant and then this repo: https://github.com/danveloper/flash-moe Innovations like these will make local hosting more and more practical.
2 likes • 21d
na.. try it out you will be waiting 20 minutes before the hello prompt responds..
0 likes • 21d
@Vamsi Acharya OK..
Gemma 4 is out with a nice Opensource Licence
https://www.youtube.com/watch?v=Kaq5Ual2ij8 @Stephen King gonna try it later, setting up something else for my local more secure info set up.
2 likes • 21d
I have Gemma running on one of my OpenCLaw instanances a MacMini with 24GB ram. It runs and is pretty solid. Don't give it any too complex tasks though but web scraping or reporting on the weather etc.. it does those simple tasks just fine.
Hermes Agent... Self Improving AI that can run local?
I'm looking at this right now. It super new, but how secure, will it go virial, will it get as much support? https://www.youtube.com/watch?v=EHlqRx0r4BI
2 likes • 24d
I installed it and it kept hanging.. reverted back to OpenClaw which is working.
Better Than a Mac Mini, Local and Private
Things got a kickstarter, but I'm good for now. Kinda wish I didn't buy the ram I needed, have to make income from this before I start buying more toys :P https://www.youtube.com/watch?v=RkzCAaIV_cQ
1 like • 25d
@Stephen King it only as 24GB of Vmem..
OLLAMA models optimized for Apple Silicon
Go this email and tried it out.. really workable and the speed is good. ollama launch claude --model qwen3.5:35b-a3b-coding-nvfp4 email from Ollama.com Today, we're previewing the fastest way to run Ollama on Apple silicon, powered by MLX, Apple's machine learning framework. This change accelerates your most demanding work on macOS: Personal assistants like OpenClaw Coding agents like Claude Code, OpenCode, or Codex Fastest performance on Apple silicon, powered by MLX Ollama on Apple silicon is now built on top of Apple’s machine learning framework, MLX, to take advantage of its unified memory architecture. This results in a large speedup of Ollama on all Apple Silicon devices. On Apple’s M5, M5 Pro and M5 Max chips, Ollama leverages the new GPU Neural Accelerators to accelerate both time to first token (TTFT) and generation speed (tokens per second). Testing was conducted using Alibaba’s Qwen3.5-35B-A3B model quantized to NVFP4 and Ollama’s previous implementation quantized to Q4_K_M using Ollama 0.18. Ollama 0.19 will see even higher performance (1851 token/s prefill and 134 token/s decode) when running withint4 quantization NVFP4 support: higher quality responses and production parity Ollama now leverages NVIDIA’s NVFP4 format to maintain model accuracy while reducing memory bandwidth and storage requirements for inference workloads. As more inference providers scale inference using NVFP4 format, this allows Ollama users to share the same results as they would in a production environment. It further opens up Ollama to have the ability to run models optimized by NVIDIA’s model optimizer. Other precisions will be made available based on the design and usage intent from Ollama’s research and hardware partners. Improved caching for more responsiveness Ollama’s cache has been upgraded to make coding and agentic tasks more efficient. Lower memory utilization: Ollama will now reuse its cache across conversations, meaning less memory utilization and more cache hits when branching when using a shared system prompt with tools like Claude Code.
2 likes • 26d
MacBook Pro G4 128GB ram
1-5 of 5
Wally Barr
2
7points to level up
@wally-barr-9198
Looking to improve the tail end of my career as an analyst. Am a family man and I travel ofter. Have background in Law Enforcement and CyberSecurity.

Active 21d ago
Joined Apr 1, 2026
Powered by