
FunAI
@FunAILab
Followers
209
Following
11
Media
1
Statuses
9
Research lab led by @y_m_asano at @utn_nuremberg. We conduct fundamental AI research and develop core technology for future Foundation Models.
Nuremberg
Joined July 2024
Now finally accepted at @emnlpmeeting! I think the technique and high-level ideas i) allow bidirectional attention for prompt & ii) (maybe) process input-query differently from answer generation will stick around.
Today we introduce Bidirectional Instruction Tuning (Bitune). It's a new way of adapting LLMs for the instruction->answering stage. It allows the model to process the instruction/question with bidirectional attention, while the answer generation remains causal.
1
4
50
Today we release Franca, a new vision Foundation Model that matches and sometimes outperforms DINOv2. The data, the training code and the model weights (with intermediate checkpoints) are open-source, allowing everyone to build on this. Methodologically, we introduce two new
Can open-data models beat DINOv2? Today we release Franca, a fully open-sourced vision foundation model. Franca with ViT-G backbone matches (and often beats) proprietary models like SigLIPv2, CLIP, DINOv2 on various benchmarks setting a new standard for open-source research๐งต
3
24
174
Hello FunAI Lab at UTN ๐ Iโm excited to start a new chapter of my research journey here in Nuremberg as a visiting postdoc. Excited for inspiring collaborations and impactful research ahead with @y_m_asano and the amazing students๐
0
7
21
LoRA et al. enable personalised model generation and serving, which is crucial as finetuned models still outperform general ones in many tasks. However, serving a base model with many LoRAs is very inefficient! Now, there's a better way: enter Prompt Generation Networks,
1
9
93
Is the community trying to surprise us today? ๐คฏ Because these benchmark-related papers from different research labs all dropped on the Daily Papers page at once! ๐๐ https://t.co/pizTMDvIGc โจ LOKI: A Comprehensive Synthetic Data Detection Benchmark using Large Multimodal
huggingface.co
0
6
24
Today, we're introducing TVBench! ๐น๐ฌ Video-language evaluation is crucial, but are we doing it right? We find that current benchmarks fall short in testing temporal understanding. ๐งต๐
2
13
68
First paper with our FunAI Lab affiliation :)
Ever wondered if better LLMs actually have a better understanding of the visual world? ๐ค As it turns out, they do! We find: An LLM's MMLU performance correlates positively with zero-shot performance in a CLIP-like case when using that LLM to encode the text. ๐งต๐
0
0
3