nicolayr_ Profile Banner
Nicolay Rusnachenko Profile
Nicolay Rusnachenko

@nicolayr_

Followers
473
Following
15K
Media
277
Statuses
4K

💼 NLP for Radiology / Healthcare ⚕️ @BU_Research・PhD in NLP・10+ years in Information Retrieval and Software Dev (https://t.co/MsXK0rEMjl)・Opinions are mine

Bournemouth / London, UK
Joined December 2015
Don't wanna be here? Send us removal request.
@nicolayr_
Nicolay Rusnachenko
1 day
🚀 Excited to share the latest video in which I review our recently published Teacher-Student framework that being applied in multilingual clinical case report sukmmarization! 🩺📋. This is 15-min skimming of the most recently submitted work in which we
Tweet media one
0
0
2
@nicolayr_
Nicolay Rusnachenko
2 hours
💎Notable set of evaluations of API providers on prompt caching , dedicated for improving LM response performance.
@chenchenygu
Chenchen Gu
24 hours
Prompt caching lowers inference costs but can leak private information from timing differences. Our audits found 7 API providers with potential leakage of user data. Caching can even leak architecture info—OpenAI's embedding model is likely a decoder-only Transformer!.🧵1/9
Tweet media one
0
0
0
@nicolayr_
Nicolay Rusnachenko
2 hours
🤔Curious how how this idea of revealing most meaningful attention heads of LLMs could be used in the analysis of the certain and domain specific tasks.
@kayo_yin
Kayo Yin ✈️ ICML
5 months
Induction heads are commonly associated with in-context learning, but are they the primary driver of ICL at scale?. We find that recently discovered "function vector" heads, which encode the ICL task, are the actual primary drivers of few-shot ICL. 🧵
Tweet media one
0
0
0
@nicolayr_
Nicolay Rusnachenko
14 hours
💎Observations on how the mention of colon ":" affects overall LLM behavior and particularly impacts judgment, is very intriguing 🤯 👀.
@omarsar0
elvis
1 day
One Token to Fool LLM-as-a-Judge. Watch out for this one, devs!. Semantically empty tokens, like “Thought process:”, “Solution”, or even just a colon “:”, can consistently trick models into giving false positive rewards. Here are my notes:
Tweet media one
0
0
0
@nicolayr_
Nicolay Rusnachenko
1 day
The review of the system is now available:. 🧵2/n.
0
0
0
@nicolayr_
Nicolay Rusnachenko
1 day
RT @aclmeeting: 🤯 Get ready for #ACL2025NLP! featuring 3500+ paper presentations (talks & posters!), numerous workshops, several tutorials….
0
23
0
@nicolayr_
Nicolay Rusnachenko
3 days
📝 Notably the problem quality reduction of LLM services formed as a chat to be used for long conversation could be referred as "context rot".
@QuentinAnthon15
Quentin Anthony
4 days
Along this point, there's a long tail of issues that cause an LLM to choke:.- "Context rot", where models become distracted by long+irrelevant contexts (especially from long conversations). See You need to open a new chat often. This effect is worsened if.
0
0
0
@nicolayr_
Nicolay Rusnachenko
3 days
RT @rasbt: Kimi K2 is basically DeepSeek V3 but with fewer heads and more experts:
Tweet media one
0
514
0
@nicolayr_
Nicolay Rusnachenko
3 days
📢 Kimi K2 is something that should not be missed out.
@huggingface
Hugging Face
3 days
Kimi K2 is number one trending on HF, congrats!
Tweet media one
0
0
2
@nicolayr_
Nicolay Rusnachenko
4 days
💎 The hierarchical structuring of the existing LLM systems suitable for the various tuning scenarios in the Healthcare NLP domain . #healthcare #nlp #llm #genai #ontology.
@omarsar0
elvis
2 years
A Survey of LLMs for Healthcare. This looks like a nice comprehensive overview of LLMs applied to the healthcare domain.
Tweet media one
0
0
0
@nicolayr_
Nicolay Rusnachenko
4 days
💎 findings on benchmarking of LLM capabilities in the domain of healthcare and information Retrieval on clinical reports / clinical notes 📊.
@emollick
Ethan Mollick
2 years
Surprisingly, a Large Language Model trained on health systems data did a better job predicting patient outcomes than traditional machine learning methods. “we show that it is possible to use LLMs as universal prediction engines for a wide range of medical predictive tasks.”
Tweet media one
0
0
2
@nicolayr_
Nicolay Rusnachenko
4 days
RT @GoogleResearch: Introducing new models for research & development of health applications: MedGemma 27B Multimodal, for complex multimod….
0
204
0
@nicolayr_
Nicolay Rusnachenko
4 days
RT @osanseviero: I'm excited to share the launch of MedGemma 💎. 🤗4B multimodal and 27B thinking text models.👀 Image classification and inte….
0
74
0
@nicolayr_
Nicolay Rusnachenko
4 days
RT @mervenoyann: MedGemma Concept Apps: MedGemma and MedSigLIP: 😍.
0
2
0
@nicolayr_
Nicolay Rusnachenko
8 days
🚨 Exciting update in healthcare AI! Thrilled to share our latest advances in Information Retrieval on shortening clinical case reports 📝🩺 . Our studies made at Bournemouth University wrapped into system submisssion "" has been accepted @ 🎤 CLEF 2025 for BioASQ Workshop
Tweet media one
1
0
1
@nicolayr_
Nicolay Rusnachenko
10 days
RT @ai_for_success: Large Language Models are improving at an exponential rate. If the pace continues until 2030, they will be able to comp….
0
201
0
@nicolayr_
Nicolay Rusnachenko
13 days
RT @omarsar0: Sometimes you get lucky with vibe coding. These days, I rely less on luck and get better results by focusing on context eng….
0
17
0
@nicolayr_
Nicolay Rusnachenko
13 days
RT @dmsobol: Thanks to @aiDotEngineer for releasing the recording of our Mixture of Agents workshop! . Watch it here: .
0
3
0
@nicolayr_
Nicolay Rusnachenko
13 days
RT @reach_vb: DAMN! DeepSeek R1T2 - 200% faster than R1-0528 & 20% faster than R1 🔥. Significantly better than R1 on GPQA & AIME 24. made v….
0
121
0
@nicolayr_
Nicolay Rusnachenko
16 days
RT @_akhaliq: MiCo. Multi-image Contrast for Reinforcement Visual Reasoning
Tweet media one
0
13
0