Taewhoo Lee Profile
Taewhoo Lee

@taewhoolee

Followers
22
Following
24
Media
6
Statuses
12

m.s. student in #NLP at Korea University

Joined September 2023
Don't wanna be here? Send us removal request.
@taewhoolee
Taewhoo Lee
3 months
Excited to present ๐„๐“๐‡๐ˆ๐‚ at #NAACL2025 ! Looking forward to open discussions on long-context modeling, evaluation, or anything else :). ๐Ÿ—“๏ธ Friday, May 2, 11:00-12:30.๐Ÿ“ Hall 3, Session K, Poster Session 8
Tweet media one
1
2
17
@taewhoolee
Taewhoo Lee
7 months
This project was done in collaboration with @cw_yoon99 @TigerKyo @DonghyeonLee_KR @_MinjuSong and @hyunjae__kim. Huge thanks for their amazing support and contributions!.
0
0
2
@taewhoolee
Taewhoo Lee
7 months
Even when using the same documents and cognitive demands, models perform consistently better on low-IC tasks compared to high-IC tasks. This suggests that information coverage is a key factor that significantly impacts model performance.
Tweet media one
1
0
0
@taewhoolee
Taewhoo Lee
7 months
Experimental results reveal significant performance drops in modern LLMs, highlighting a critical challenge in managing long contexts.
Tweet media one
1
0
0
@taewhoolee
Taewhoo Lee
7 months
To address this issue, we propose ๐„๐“๐‡๐ˆ๐‚, a suite of.long-context tasks specifically designed to assess.whether LLMs can fully utilize the provided information. ๐„๐“๐‡๐ˆ๐‚ comprises four tasks with different cognitive demands, and spans across four domains.
Tweet media one
1
0
0
@taewhoolee
Taewhoo Lee
7 months
We first introduce Information Coverage (IC), a new metric that measures the proportion of input context required to answer a query. We find that existing benchmarks exhibit low IC, indicating that models can solve them without fully digesting the given context.
Tweet media one
1
0
0
@taewhoolee
Taewhoo Lee
7 months
Research on LLM context windows has advanced rapidly, accompanied by the development of diverse long-context benchmarks. However, is it fair to say that a model can fully digest, e.g. 32k tokens, if the answer can be derived from a single sentence or paragraph?.
1
0
0
@taewhoolee
Taewhoo Lee
7 months
๐Ÿค”Modern LLMs are known to support long text, but can they ๐Ÿ๐ฎ๐ฅ๐ฅ๐ฒ ๐ฎ๐ญ๐ข๐ฅ๐ข๐ณ๐ž the information available in these texts?. ๐Ÿ’กIntroducing ๐„๐“๐‡๐ˆ๐‚, a new long-context benchmark designed to assess LLMs' ability to leverage the entire given context.
Tweet card summary image
arxiv.org
Recent advancements in large language models (LLM) capable of processing extremely long texts highlight the need for a dedicated evaluation benchmark to assess their long-context capabilities....
1
5
9
@taewhoolee
Taewhoo Lee
9 months
Had so much fun attending #EMNLP2024 ! Every conversation I had with fellow researchers was truly inspiring and insightful. Big thanks to @cw_yoon99 @HyeonHwang8 @jeongminby98858 for the amazing teamwork over the past several months. Moving on to the next one!
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
1
14
@taewhoolee
Taewhoo Lee
11 months
Happy to share that CompAct has been accepted to EMNLP 2024 Main ๐ŸŽ‰. Congratulations to our team @cw_yoon99 @HyeonHwang8 @jeongminby98858 , and see you in Miami !.
@cw_yoon99
Chanwoong Yoon
1 year
๐Ÿš€Looking for an advanced compressor for multi-hop QA tasks while leveraging increased top-k documents effectively? Introducing โœจCompActโœจ, a novel framework that employs an active strategy for compressing extensive documents. [1/5]. paper:
0
0
9
@taewhoolee
Taewhoo Lee
2 years
RT @karpathy: # On the "hallucination problem". I always struggle a bit with I'm asked about the "hallucination problem" in LLMs. Because,โ€ฆ.
0
3K
0