
Nicolay Rusnachenko
@nicolayr_
Followers
473
Following
15K
Media
277
Statuses
4K
💼 NLP for Radiology / Healthcare ⚕️ @BU_Research・PhD in NLP・10+ years in Information Retrieval and Software Dev (https://t.co/MsXK0rEMjl)・Opinions are mine
Bournemouth / London, UK
Joined December 2015
💎Notable set of evaluations of API providers on prompt caching , dedicated for improving LM response performance.
Prompt caching lowers inference costs but can leak private information from timing differences. Our audits found 7 API providers with potential leakage of user data. Caching can even leak architecture info—OpenAI's embedding model is likely a decoder-only Transformer!.🧵1/9
0
0
0
🤔Curious how how this idea of revealing most meaningful attention heads of LLMs could be used in the analysis of the certain and domain specific tasks.
Induction heads are commonly associated with in-context learning, but are they the primary driver of ICL at scale?. We find that recently discovered "function vector" heads, which encode the ICL task, are the actual primary drivers of few-shot ICL. 🧵
0
0
0
💎Observations on how the mention of colon ":" affects overall LLM behavior and particularly impacts judgment, is very intriguing 🤯 👀.
One Token to Fool LLM-as-a-Judge. Watch out for this one, devs!. Semantically empty tokens, like “Thought process:”, “Solution”, or even just a colon “:”, can consistently trick models into giving false positive rewards. Here are my notes:
0
0
0
RT @aclmeeting: 🤯 Get ready for #ACL2025NLP! featuring 3500+ paper presentations (talks & posters!), numerous workshops, several tutorials….
0
23
0
📝 Notably the problem quality reduction of LLM services formed as a chat to be used for long conversation could be referred as "context rot".
Along this point, there's a long tail of issues that cause an LLM to choke:.- "Context rot", where models become distracted by long+irrelevant contexts (especially from long conversations). See You need to open a new chat often. This effect is worsened if.
0
0
0
💎 The hierarchical structuring of the existing LLM systems suitable for the various tuning scenarios in the Healthcare NLP domain . #healthcare #nlp #llm #genai #ontology.
A Survey of LLMs for Healthcare. This looks like a nice comprehensive overview of LLMs applied to the healthcare domain.
0
0
0
💎 findings on benchmarking of LLM capabilities in the domain of healthcare and information Retrieval on clinical reports / clinical notes 📊.
Surprisingly, a Large Language Model trained on health systems data did a better job predicting patient outcomes than traditional machine learning methods. “we show that it is possible to use LLMs as universal prediction engines for a wide range of medical predictive tasks.”
0
0
2
RT @GoogleResearch: Introducing new models for research & development of health applications: MedGemma 27B Multimodal, for complex multimod….
0
204
0
RT @osanseviero: I'm excited to share the launch of MedGemma 💎. 🤗4B multimodal and 27B thinking text models.👀 Image classification and inte….
0
74
0
RT @ai_for_success: Large Language Models are improving at an exponential rate. If the pace continues until 2030, they will be able to comp….
0
201
0
RT @omarsar0: Sometimes you get lucky with vibe coding. These days, I rely less on luck and get better results by focusing on context eng….
0
17
0
RT @dmsobol: Thanks to @aiDotEngineer for releasing the recording of our Mixture of Agents workshop! . Watch it here: .
0
3
0
RT @reach_vb: DAMN! DeepSeek R1T2 - 200% faster than R1-0528 & 20% faster than R1 🔥. Significantly better than R1 on GPQA & AIME 24. made v….
0
121
0