mark_l_watson Profile Banner
mark_l_watson Profile
mark_l_watson

@mark_l_watson

Followers
3K
Following
7K
Media
41
Statuses
3K

AI Generalist. Author 20 books on AI, LLMs, DL, semantic web, Lisp. 55 patents. My recent books are free to read or buy online https://t.co/aw3vKvefPe

Flagstaff Arizona
Joined October 2008
Don't wanna be here? Send us removal request.
@mark_l_watson
mark_l_watson
4 hours
Gemini 3 Flash, at $0.50 per million input tokens, hits a sweet spot for performance and low cost. I don’t know yet, but I might not have to use the Pro API, which is 4 times as expensive, very often.
0
0
0
@mark_l_watson
mark_l_watson
2 days
The new 30B model nvidia/nemotron-3-nano is very good with tool use - I just checked it using LM Studio against all of my tool and localMCP server projects. A pleasant surprise: I have a 32G M2-Pro Mac system, and the model weights are over 20G, but performance is quite good -
0
0
4
@mark_l_watson
mark_l_watson
2 days
I think Gemini will overtake ChatGPT. My non-tech brother spent a few minutes on the phone yesterday raving about how he uses Gemini on his new Pixel phone for just about everything. I am curious to see how good the Gemini iOS integration will be next year. One more advantage: I
@mark_k
Mark Kretschmann
3 days
Do you think Gemini will become the default AI option in the future rather than ChatGPT?
0
1
4
@mark_l_watson
mark_l_watson
2 days
After Wave was dropped, I still played around with the open source release for a year - a major waste of time and I had problems understanding and slightly modifying the code base, but I had a lot of fun with it, so not a flop for me! re: grandparent comment about AntiGravity
@codymclain
Cody McLain
4 days
@AdityaShips idk, I've gotta say Google Wave was a bigger flop. It tried to replace email and ended up confusing everyone. Sometimes tech tries to do too much and ends up lost in translation
0
0
1
@mark_l_watson
mark_l_watson
4 days
I have a project that I want to work on next year - my goal is not to build anything practical, rather to experiment with ideas. I am trying to decide whether to use Python, Common Lisp, or Clojure. I stood up a project web site as a placeholder:
0
0
6
@mark_l_watson
mark_l_watson
4 days
At the command line, I use (ql:quickload …) 95% of the time and (load …) 5% of the time. In my .sbclrc file I alias ‘ql’ to ‘ql:quickload’. In Emacs I open a REPL and manually (ql ….) a project, then just evaluate expressions as I
@killian_arts
KILLIAN.arts
5 days
I have some questions for my Common Lisp bros about this blog going around: https://t.co/TIs1SJbUSY How often do you use LOAD or COMPILE-FILE, rather than some slime or sly command in Emacs? And do you prefer UIOP:DEFINE-PACKAGE or the standard DEFPACKAGE?
1
0
9
@mark_l_watson
mark_l_watson
5 days
OK, running Mistral vide with local ‘ollama run devstral-small-2’ is really too slow (but works OK). Running ollama on command line with a detailed prompt and copying code to be modified is more practical:
@mark_l_watson
mark_l_watson
5 days
Hey! devstral-small-2 is a very good coding model. I don’t have it hooked up to Mistral Vibe yet, just running with ‘ollama run devstral-small-2’ copying in some of my older Clojure and Python code and asking for improvements. All modified code ran tests correctly. Amazing for a
0
1
2
@mark_l_watson
mark_l_watson
5 days
Hey! devstral-small-2 is a very good coding model. I don’t have it hooked up to Mistral Vibe yet, just running with ‘ollama run devstral-small-2’ copying in some of my older Clojure and Python code and asking for improvements. All modified code ran tests correctly. Amazing for a
0
0
4
@mark_l_watson
mark_l_watson
7 days
For a dense 8B model rnj-1 runs quickly on my Mac. I experiment with my own low level (no dependencies) agentic code so fast local execution is important:
@mark_l_watson
mark_l_watson
7 days
So nice: rnj-1 from essentialAI. I have tested it on a variety of tool use cases and asked for Python code generation for a graph theory problem. 100% pass on my tool use test cases and the 100 lines of generated Python code worked perfectly. Run with: 'ollama run rnj-1’ or use
1
0
1
@mark_l_watson
mark_l_watson
7 days
So nice: rnj-1 from essentialAI. I have tested it on a variety of tool use cases and asked for Python code generation for a graph theory problem. 100% pass on my tool use test cases and the 100 lines of generated Python code worked perfectly. Run with: 'ollama run rnj-1’ or use
0
0
1
@mark_l_watson
mark_l_watson
13 days
Some more information: One working Mistrail-3:3b tool example uses the standard Ollama Python setup (tools=[local_files] inside https://t.co/ckgdiUgB8W). My second example uses tools from scratch using only the Python requests library:
@mark_l_watson
mark_l_watson
13 days
I used the Ollama Python library for:
0
0
1
@mark_l_watson
mark_l_watson
13 days
I used the Ollama Python library for:
@gregbarbosa
Greg Barbosa
13 days
@mark_l_watson what are you using to get 3B for tool use? I haven't been able to get it to work in LM Studio yet. Might try llama.cpp next
0
0
0
@mark_l_watson
mark_l_watson
14 days
I enjoy writing code from scratch, especially when experimenting with tool use and LLM-based agents. I am pleasantly surprised at how well ministral-3:3b does for tool use - probably a priority training the model.
1
0
4
@QuixiAI
Eric Hartford
15 days
Bravo @MistralAI ! Thank you for the Apache 2.0 license! And the MoE and the multimodal! Everything I've been wishing for is delivered. A thinking multimodal vision model! I will start mapping the refusal space. (Compared with Qwen3)
4
7
132
@natolambert
Nathan Lambert
15 days
Mistral is back to their roots All apache 2.0 Strong models, multimodal, multiple types, and MORE BASE MODELS Huge W, smart by them to capitalize on the power vacuum of companies looking for alternates to Chinese models. So happy Chinese companies are pushing change globally.
@MistralAI
Mistral AI
15 days
Introducing the Mistral 3 family of models: Frontier intelligence at all sizes. Apache 2.0. Details in 🧵
27
64
965
@mark_l_watson
mark_l_watson
14 days
For a 8B parameter model mistral-3:8b seems very good, testing it now with my collection of 'oddball' private test prompts. I have seen some negative comments on the new Mistral releases but for small embedded and local applications Mistral is in the same ballpark as other 2025
0
0
1
@mark_l_watson
mark_l_watson
21 days
Gary, I enjoyed your last book and just signed up for your newsletter. I for one am happy to see more basic non-LLM research (even though I spend almost an hour a day using LLM based systems as tools).
@GaryMarcus
Gary Marcus
21 days
This place is toxic. For the last seven years I warned you that LLMs and similar approaches would not lead us to AGI. Almost nobody is willing to acknowledge that, even though so many of you gave me endless grief about it at the time. I also warned you -– first –- that Sam
0
0
6