solyarisoftware Profile Banner
Giorgio Robino Profile
Giorgio Robino

@solyarisoftware

Followers
3K
Following
15K
Media
1K
Statuses
29K

Conversational LLM-based Applications Specialist @almawave | Former ITD-CNR Researcher | Soundscapes (Orchestral) Composer.

Genova, Italia
Joined April 2009
Don't wanna be here? Send us removal request.
@solyarisoftware
Giorgio Robino
10 months
My preprint "Conversation Routines: A Prompt Engineering Framework for Task-Oriented Dialog Systems" now has a revised version on @arXiv with updated experimental results. Here’s a thread with the changes! 🧡 ➑️ Paper: https://t.co/8kMIiB5zbu 1/ What’s CR?
Tweet card summary image
arxiv.org
This study introduces Conversation Routines (CR), a structured prompt engineering framework for developing task-oriented dialog systems using Large Language Models (LLMs). While LLMs demonstrate...
1
2
3
@askOkara
Okara
3 days
GLM-4.6V is insanely good at frontend tasks πŸ”₯ Give it a screenshot or design file and it detects layouts, components and color schemes then generates high-fidelity HTML/CSS/JS code. Try it now on https://t.co/M5laJrXX0V
3
11
133
@manthanguptaa
Manthan Gupta
2 days
I spent the last few days prompting ChatGPT to understand how its memory system actually works. Spoiler alert: There is no RAG used https://t.co/zxvRRP2GK8
224
459
4K
@rohanpaul_ai
Rohan Paul
2 days
New OpenAI paper Shows how to train a language model to add an honest "confession" after its normal answer. The work starts from the concern that large models can bluff, hide mistakes, or quietly game the reward signals used in training. To address this, the authors add a
10
19
102
@rohanpaul_ai
Rohan Paul
2 days
The paper shows how a small language model can truly think in parallel instead of step by step. Normal models generate 1 long chain of reasoning, explore only 1 path, and waste time when that path is wrong. Native Parallel Reasoner is a method that lets the model split a
5
37
248
@jiqizhixin
ζœΊε™¨δΉ‹εΏƒ JIQIZHIXIN
1 day
Google just found the agentic scaling law! Forget "more agents is all you need." After 180 experiments across GPT, Gemini, and Claude, the results are in: - The 45% Trap: If a single agent has >45% accuracy, adding more agents often hurts performance. - Tool Tax: Tool-heavy
21
127
1K
@Hesamation
ℏΡsam
10 days
this is big... 50 AI researchers from Bytedance, Alibaba, Tencent, and other labs/universities just published a 300-page paper with surprising lessons about coding models and agents (data, pre and post-training, etc). key highlights: > small LLMs can beat proprietary giants RL
39
357
2K
@qizhengz_alex
Qizheng Zhang
10 days
πŸš€ Introducing Agentic Context Engineering (ACE) --- a framework for self-improving language models through continuously evolving contexts (not weights). πŸ“ˆ High-Performing: +10.6% on agent tasks, +8.6% on finance ⚑ Ultra-Efficient: βˆ’86.9% latency, βˆ’83.6% dollar cost πŸ› 
23
114
753
@maximelabonne
Maxime Labonne
12 days
ToolOrchestra is such a cool work from @nvidia Just an 8B model trained on calling tools and other LLMs to answer queries It's a great demo of what frontier SLMs will be about in 2026
15
79
488
@rohanpaul_ai
Rohan Paul
15 days
The paper shows how to let vision language models think with compact visual thoughts inside the model, not just with words. Current models turn an image into a few text features, so they lose exact object boundaries, distances, and geometry, which makes them weak at tasks like
5
76
352
@Hesamation
ℏΡsam
12 days
context engineering is an essential skill and surprisingly simple to learn. a free 40-page guide just dropped, with amazing diagrams to intuitively teach you about: > how to architecture your agents > retrieval, chunking techniques > prompting strategies > architecture of agentic
13
74
568
@Alibaba_Qwen
Qwen
11 days
llama.cpp (PR #16095) just added support for Qwen3-Next β€” Qwen’s new hybrid architecture! You can now run Qwen3-Next locally with efficient CPU/GPU inference. πŸš€ πŸ”— PR:
Tweet card summary image
github.com
EDIT: README FIRST This is an implementation of a new type of attention gating in GGML. Therefore, this implementation will be focused on CORRECTNESS ONLY. Speed tuning and support for more archite...
38
112
888
@askalphaxiv
alphaXiv
12 days
How to Properly do LLM-as-a-Judge Raw LLM-as-a-Judge scores are inherently biased due to how LLMs would often make mistakes This paper proposes a simple statistical method to correct the scores and calculate valid confidence intervals via a human-verified calibration set
16
89
603
@JustinLin610
Junyang Lin
15 days
preparing for the next generation model and the vision should be a very important pary of it. techniques applied in qwen3-vl are mostly quite effective and we r now moving towards more multimodal native training, data scaling of tasks and domains in vision, and multimodal agentic
@Alibaba_Qwen
Qwen
15 days
πŸš€ Qwen3-VL Tech report is now out on arXiv! From pretraining to post-training, architecture to infra, data to evaluation β€” we’ve packed in the details for anyone building on vision-language models. πŸ”₯ 3 models >1M downloads in just over a month πŸ† Qwen3-VL-8B leads with 2M+
23
32
509
@karminski3
karminski-η‰™εŒ»
13 days
Qwen3-Next-80B-A3B ηš„ unsloth η‰ˆζœ¬ gguf η»ˆδΊŽε‘εΈƒε•¦οΌθΏ™δΈͺζ¨‘εž‹η»ˆδΊŽε―δ»₯εœ¨ζœ¬εœ°η—›εΏ«ηš„η”¨δΊ†γ€‚ζˆ‘128G ηš„M2 Ultra ηŽ°εœ¨θ·‘6bit ιžεΈΈδΈζ»‘. ηŽ°εœ¨ε°±η­‰δΈ€ζ³’εΉ΄εΊ•ηš„ Next
11
26
143
@MaryamMiradi
Maryam Miradi, PhD
14 days
Why Your LLM Needs an "Off Switch": π—”π˜π˜π—²π—»π˜π—Άπ—Όπ—» 𝗦𝗢𝗻𝗸 Congratulations to Qwen for winning the NeurIPS 2025 Best Paper Award! Attention sinks are finally getting the attention they deserve. Here's what most people miss about modern LLMs: Almost 80% of attention in
1
5
15
@NielsRogge
Niels Rogge
14 days
Nvidia silently dropped Orchestrator-8B πŸ‘€ β€œOn the Humanity's Last Exam (HLE) benchmark, ToolOrchestrator-8B achieves a score of 37.1%, outperforming GPT-5 (35.1%) while being approximately 2.5x more efficient.”
Tweet card summary image
huggingface.co
34
151
1K
@rohanpaul_ai
Rohan Paul
14 days
New Apple paper makes RAG shorter and smarter by training retrieval and generation together in one shared continuous space. Standard RAG picks documents using separate embeddings and then feeds long raw text to the model, so retrieval and answer quality cannot influence each
8
34
180
@StepFun_ai
StepFun
15 days
πŸš€ Introducing Step-Audio-R1: The first audio LLM to unlock test-time compute scaling! πŸ› οΈ Key Features: - Deep Audio Comprehension - Real-time responsiveness - Scalable chain-of-thought reasoning for audio tasks πŸ”₯ Performance: βœ… Surpasses Gemini 2.5 Pro & comparable to
Tweet card summary image
huggingface.co
10
83
589
@dair_ai
DAIR.AI
15 days
Banger paper from NVIDIA. Bigger models aren't always the answer. However, the default approach to improving AI systems today remains scaling up. More parameters, more compute, more cost. But many tasks don't require the full power of a massive model. This new research
20
129
760
@e_opore
Dhanian πŸ—―οΈ
14 days
How AI Agents Use Memory Systems 1. Introduction β†’ Memory is essential for AI agents because it allows them to retain information, reason across time, and improve decisions based on past interactions. β†’ Without memory, agents would act blindly, unable to learn or adapt. 2.
22
92
465