mark_l_watson
@mark_l_watson
Followers
3K
Following
7K
Media
41
Statuses
3K
AI Generalist. Author 20 books on AI, LLMs, DL, semantic web, Lisp. 55 patents. My recent books are free to read or buy online https://t.co/aw3vKvefPe
Flagstaff Arizona
Joined October 2008
Gemini 3 Flash, at $0.50 per million input tokens, hits a sweet spot for performance and low cost. I don’t know yet, but I might not have to use the Pro API, which is 4 times as expensive, very often.
0
0
0
The new 30B model nvidia/nemotron-3-nano is very good with tool use - I just checked it using LM Studio against all of my tool and localMCP server projects. A pleasant surprise: I have a 32G M2-Pro Mac system, and the model weights are over 20G, but performance is quite good -
0
0
4
I think Gemini will overtake ChatGPT. My non-tech brother spent a few minutes on the phone yesterday raving about how he uses Gemini on his new Pixel phone for just about everything. I am curious to see how good the Gemini iOS integration will be next year. One more advantage: I
0
1
4
After Wave was dropped, I still played around with the open source release for a year - a major waste of time and I had problems understanding and slightly modifying the code base, but I had a lot of fun with it, so not a flop for me! re: grandparent comment about AntiGravity
@AdityaShips idk, I've gotta say Google Wave was a bigger flop. It tried to replace email and ended up confusing everyone. Sometimes tech tries to do too much and ends up lost in translation
0
0
1
I have a project that I want to work on next year - my goal is not to build anything practical, rather to experiment with ideas. I am trying to decide whether to use Python, Common Lisp, or Clojure. I stood up a project web site as a placeholder:
0
0
6
At the command line, I use (ql:quickload …) 95% of the time and (load …) 5% of the time. In my .sbclrc file I alias ‘ql’ to ‘ql:quickload’. In Emacs I open a REPL and manually (ql ….) a project, then just evaluate expressions as I
I have some questions for my Common Lisp bros about this blog going around: https://t.co/TIs1SJbUSY How often do you use LOAD or COMPILE-FILE, rather than some slime or sly command in Emacs? And do you prefer UIOP:DEFINE-PACKAGE or the standard DEFPACKAGE?
1
0
9
OK, running Mistral vide with local ‘ollama run devstral-small-2’ is really too slow (but works OK). Running ollama on command line with a detailed prompt and copying code to be modified is more practical:
Hey! devstral-small-2 is a very good coding model. I don’t have it hooked up to Mistral Vibe yet, just running with ‘ollama run devstral-small-2’ copying in some of my older Clojure and Python code and asking for improvements. All modified code ran tests correctly. Amazing for a
0
1
2
Hey! devstral-small-2 is a very good coding model. I don’t have it hooked up to Mistral Vibe yet, just running with ‘ollama run devstral-small-2’ copying in some of my older Clojure and Python code and asking for improvements. All modified code ran tests correctly. Amazing for a
0
0
4
Well, I feel better informed now than I did an hour ago before I read @Tim_Dettmers article:
timdettmers.com
If you are reading this, you probably have strong opinions about AGI, superintelligence, and the future of AI. Maybe you believe we are on the cusp of a transformative breakthrough. Maybe you are...
3
1
10
For a dense 8B model rnj-1 runs quickly on my Mac. I experiment with my own low level (no dependencies) agentic code so fast local execution is important:
So nice: rnj-1 from essentialAI. I have tested it on a variety of tool use cases and asked for Python code generation for a graph theory problem. 100% pass on my tool use test cases and the 100 lines of generated Python code worked perfectly. Run with: 'ollama run rnj-1’ or use
1
0
1
So nice: rnj-1 from essentialAI. I have tested it on a variety of tool use cases and asked for Python code generation for a graph theory problem. 100% pass on my tool use test cases and the 100 lines of generated Python code worked perfectly. Run with: 'ollama run rnj-1’ or use
0
0
1
Some more information: One working Mistrail-3:3b tool example uses the standard Ollama Python setup (tools=[local_files] inside https://t.co/ckgdiUgB8W). My second example uses tools from scratch using only the Python requests library:
0
0
1
I used the Ollama Python library for:
@mark_l_watson what are you using to get 3B for tool use? I haven't been able to get it to work in LM Studio yet. Might try llama.cpp next
0
0
0
I enjoy writing code from scratch, especially when experimenting with tool use and LLM-based agents. I am pleasantly surprised at how well ministral-3:3b does for tool use - probably a priority training the model.
1
0
4
Bravo @MistralAI ! Thank you for the Apache 2.0 license! And the MoE and the multimodal! Everything I've been wishing for is delivered. A thinking multimodal vision model! I will start mapping the refusal space. (Compared with Qwen3)
4
7
132
Mistral is back to their roots All apache 2.0 Strong models, multimodal, multiple types, and MORE BASE MODELS Huge W, smart by them to capitalize on the power vacuum of companies looking for alternates to Chinese models. So happy Chinese companies are pushing change globally.
Introducing the Mistral 3 family of models: Frontier intelligence at all sizes. Apache 2.0. Details in 🧵
27
64
965
For a 8B parameter model mistral-3:8b seems very good, testing it now with my collection of 'oddball' private test prompts. I have seen some negative comments on the new Mistral releases but for small embedded and local applications Mistral is in the same ballpark as other 2025
0
0
1
Gary, I enjoyed your last book and just signed up for your newsletter. I for one am happy to see more basic non-LLM research (even though I spend almost an hour a day using LLM based systems as tools).
This place is toxic. For the last seven years I warned you that LLMs and similar approaches would not lead us to AGI. Almost nobody is willing to acknowledge that, even though so many of you gave me endless grief about it at the time. I also warned you -– first –- that Sam
0
0
6