ToC_lab
@ToC_lab
Followers
9
Following
5
Media
16
Statuses
32
Theory of Computation Lab, YONSEI University
Seoul
Joined January 2025
Thank you for reading. For more details, please check out our paper: https://t.co/yzbBa0HobA (5/5)
arxiv.org
Large language models (LLMs) show strong performance across natural language processing (NLP), mathematical reasoning, and programming, and recent large reasoning models (LRMs) further emphasize...
0
0
0
This work highlights a critical gap in the spatial reasoning capabilities of current models, identifying key failure modes like repetition. We provide a rigorous new framework for future research. (4/5)
1
0
0
Our benchmark uses two formal regex tasks: Minimization & Equivalence. The empirical results are striking—even LLMs and Reasoning Models (LRMs) show a significant performance drop, especially on minimization, where most fail to produce even an equivalent expression.(3/5)
1
0
0
How do LLMs perform when faced with problems requiring massive search space exploration under memory constraints? We introduce RegexPSPACE, the first benchmark designed to test this, moving beyond NP-hard tasks to the more challenging PSPACE-complete class. (2/5)
1
0
0
📣 Excited to share our new paper "RegexPSPACE: A Benchmark for Evaluating LLM Reasoning on PSPACE-complete Regex Problems"! (1/5) Paper: https://t.co/A3asGnRYzU Code: https://t.co/jTbNyqw7BN
#AIResearch #Regex #PSPACE-complete #Benchmark #ReasoningTasks
1
0
0
🚀Just released! We will introduce "TCProF: Time-Complexity Prediction SSL Framework" at #NAACL2025. Ideal for programming competitions and code education! 🎉💻 Paper: https://t.co/fiOhQ44e8q Code: https://t.co/DWJIqJwhhc
#AIResearch #CodeTimeComplexity #SemiSupervisedLearning
4
0
2
🔍 Curious about improving implicit hate speech detection just through data selection? Our paper introduces CONELA, showing how selecting training data based on human agreement patterns & model dynamics can significantly boost performance - without changing model architecture!
0
0
1
📣 Excited to share our new paper "Analyzing Offensive Language Dataset Insights from Training Dynamics and Human Agreement Level" at #COLING2025! 🎉 Paper: https://t.co/y0kog2RdSa CONELA Code: https://t.co/mEWHMGBKeq
#NLP2025 #dk_search_ai #finetuning #hatespeechdetection
2
0
4
That’s the dataset in a nutshell! This work created by me (Seung-Yeop Baik) with the help of Joonghyuk Hahn, Jungin Kim, Mingi Jeon, Aditi, Yo-Sub Han, and Sang-Ki Ko.If you want to access the dataset or get into further details, follow the links below to get the full picture!
1
0
5
To avoid misleading, the GPT in the figure is an accusation module for the human annotators to think over their reasoning.
1
0
2
We share the annotation process of creating the dataset. CodeComplex leverages competitive programming submissions from Codeforces and builds on the CodeContests dataset developed by DeepMind.
1
0
2