_June1126 Profile Banner
Jing Xiong Profile
Jing Xiong

@_June1126

Followers
36
Following
13
Media
16
Statuses
27

Phd student in HKU. Research Direction: Efficient Natural Language Processing and Automated Theorem Proving

Joined March 2016
Don't wanna be here? Send us removal request.
@_June1126
Jing Xiong
5 days
RT @_reachsumit: CTR-Sink: Attention Sink for Language Models in Click-Through Rate Prediction. Introduces behavior-level attention sinks t….
Tweet card summary image
arxiv.org
Click-Through Rate (CTR) prediction, a core task in recommendation systems, estimates user click likelihood using historical behavioral data. Modeling user behavior sequences as text to leverage...
0
1
0
@_June1126
Jing Xiong
2 months
RT @_reachsumit: UncertaintyRAG: Span-Level Uncertainty Enhanced Long-Context Modeling for Retrieval-Augmented Generation. Uses SNR-based s….
0
5
0
@grok
Grok
12 hours
"A girl in a flowing white dress floating gracefully into a dreamy sky filled with stars and colorful clouds at sunset.". Try Grok Imagine, free for a limited time:.
12
5
57
@_June1126
Jing Xiong
2 months
#ICML2025 #ParrallelComp #Long-context #Length Extrapolation #Memory-bound #Efficient-inference #KV cache Compression #128K Token.
0
0
0
@_June1126
Jing Xiong
2 months
🚀 Our 8B LLM achieves 91.17% of GPT-4's performance on ultra-long context reasoning, surpassing formidable models such as Claude-2 and Kimi-Chat—all with only 8K context training.
Tweet media one
0
0
0
@_June1126
Jing Xiong
2 months
🧠 A key contribution is our theoretical and empirical analysis of attention bias under parallel attention. We uncover how and why attention sinks emerge and provide effective calibration strategies.
Tweet media one
0
0
0
@_June1126
Jing Xiong
2 months
🔍 We tackle memory limitations in length extrapolation by introducing parallel attention, KV cache compression, and chunk eviction strategies that break the GPU memory bottleneck—without any retraining required.
Tweet media one
0
0
0
@_June1126
Jing Xiong
2 months
Our paper has been accepted to ICML 2025! 🎉. 📢 In this paper, we propose ParallelComp, a training-free method to enable LLMs to extrapolate context length from 8K up to 128K tokens on a single A100 GPU, with minimal performance loss.
Tweet media one
0
0
0
@_June1126
Jing Xiong
2 months
🔬 The HKU team presents ParallelComp: a training-free technique for efficient context length extrapolation in LLMs—from 8K up to 128K tokens—on a single A100 GPU, with minimal performance loss. 📄 Paper: 💻 Code:
Tweet media one
Tweet media two
5
9
14
@_June1126
Jing Xiong
3 months
RT @HuiShen_umich: 📷 New Benchmark Release: PhyX - Physical Reasoning for Multimodal Models. 👉 Project Page: 👉 Gith….
0
7
0
@_June1126
Jing Xiong
11 months
RT @clin_tian: 🔥Thrilled to announce our Oral acceptance at #NeurIPS2024! 🚀HydraLoRA, an asymmetric LoRA architecture with a shared A matri….
0
14
0
@_June1126
Jing Xiong
1 year
RT @cerana99x: 🌟Excited to share LeCo's acceptance at #COLM2024! .🤔Fed up with LLMs' self-correct struggles and endless prompts?.🪄LeCo uses….
0
14
0
@_June1126
Jing Xiong
1 year
RT @ZhijiangG: +👋LLMs work quite well on modeling/understanding long context. What about generating long content 🤔. Check our ACL paper P….
Tweet card summary image
arxiv.org
Large Language Models (LLMs) have succeeded remarkably in understanding long-form contents. However, exploring their capability for generating long-form contents, such as reports and articles, has...
0
9
0
@_June1126
Jing Xiong
1 year
RT @YinhongLiu2: 🔥New paper!📜.Struggle to align LLM evaluators with human judgements?🤔.Introducing PairS🌟: By exploiting transitivity, we p….
0
10
0
@_June1126
Jing Xiong
1 year
RT @space_discrete: 翻到一篇文章[ICLR'24]Understanding Addition in Transformers.回忆起在初学oi年代被老师问了一道题:怎么直接按从左到右的顺序直接做大整数加法,不允许读完再翻转。当时想了十分钟想到一个存9和进位….
0
14
0
@_June1126
Jing Xiong
1 year
RT @_June1126: Excited to announce our paper's acceptance at ICLR 2024! 🌟 Our algorithm leverages CoT for enhanced in-context exemplar sele….
0
4
0
@_June1126
Jing Xiong
1 year
🔗 For more exciting discoveries and in-depth analysis, please check out our paper and code! #NLP #AIResearch #LanguageModels #InContextLearning #DQLoRe 📚✨
Tweet media one
Tweet media two
0
0
1
@_June1126
Jing Xiong
1 year
📊 Our experimental results showcase the exceptional performance of DQ-LoRe in multi-step reasoning tasks, especially its robustness and adaptability in distribution shift scenarios, paving new possibilities for the future application of LLMs. 🌟
Tweet media one
1
0
0
@_June1126
Jing Xiong
1 year
🔍By using the Gaussian kernel, we preserved key Chain-of-Thought info for commonsense reasoning tasks, distinguishing relevant exemplars from those similar by word co-occurrence, refining exemplar selection. 🤯
Tweet media one
1
0
0
@_June1126
Jing Xiong
1 year
Utilizing PCA dimensionality reduction, we uncover a major finding in exemplar selection: removing redundant information not only speeds up the process but also improves outcomes, resulting in a more uniform and distinguishable distribution of exemplars in the vector space.
Tweet media one
1
0
0
@_June1126
Jing Xiong
1 year
Our latest research introduces the DQ-LoRe framework, which, by combining dual queries and low-rank approximation re-ranking, significantly enhances the accuracy of exemplar selection in in-context learning. 🧠
Tweet media one
1
0
0