Daniel Mas Montserrat
@_danielmas
Followers
495
Following
4K
Media
37
Statuses
173
Building AI at @GalateaBio @Stanford đź§®
Joined March 2018
Excited to share our preprint on iLTM: an Integrated Large Tabular Model! arxiv: https://t.co/oTGRYGy3wy No single technique consistently excels across all tabular tasks. iLTM addresses this by integrating distinct paradigms in a single architecture: - Gradient Boosted
1
6
9
Despite being meta-trained exclusively on classification, iLTM transfers effectively to regression tasks with light fine-tuning, matching or surpassing strong baselines on both tasks. iLTM achieves top rankings on TabZilla Hard, TabReD, and more benchmarks, outperforming
1
1
4
Excited to share our preprint on iLTM: an Integrated Large Tabular Model! arxiv: https://t.co/oTGRYGy3wy No single technique consistently excels across all tabular tasks. iLTM addresses this by integrating distinct paradigms in a single architecture: - Gradient Boosted
1
6
9
We’re releasing code + pre-trained weights so anyone working with large-scale tabular data can get stronger baselines and build on iLTM. We’d love feedback and comparisons on your own datasets: Paper: https://t.co/oTGRYGy3wy Code: https://t.co/WLMhfKr73y Weights:
github.com
iLTM: Integrated Large Tabular Model. Contribute to AI-sandbox/iLTM development by creating an account on GitHub.
0
0
5
Unlike standard neural network, iLTM's hypernetwork explicitly parametrizes the relationship between dataset features and network weights. This allows us to visualize how the model “understands” different tasks, for example, clustering similar datasets and learning diverse
1
0
4
Despite being meta-trained exclusively on classification, iLTM transfers effectively to regression tasks with light fine-tuning, matching or surpassing strong baselines on both tasks. iLTM achieves top rankings on TabZilla Hard, TabReD, and more benchmarks, outperforming
1
1
4
iLTM is meta-trained on >1,800 heterogeneous real-world classification datasets. Instead of training from scratch for every new table, the hypernetwork learns to generate dataset-specific weights of a neural network, which can then be optionally fine-tuned and ensembled. (2/N)
1
0
5
How was this paper even accepted to ICLR? The commercial promoters of TabPFN are now trying to discredit one of the best open repositories, OpenML. Utterly unacceptable, how did this paper pass ethics board at ICLR?
1
3
7
Excited to share our latest PRS work! Our @GalateaBio and @genomelink team performed a comprehensive analysis of published @PGSCatalog models along with locally trained models using LDPred2, PRS-CSx, and SNPnet, across diverse populations using @UKBIOBANK and our own data
Polygenic risk score portability for common diseases across genetically diverse populations https://t.co/Mgw0jjT7Tf
#medRxiv
0
3
8
Polygenic risk score portability for common diseases across genetically diverse populations https://t.co/Mgw0jjT7Tf
#medRxiv
0
3
7
No son of a construction worker is just going to randomly start doing ML research if they never hear of it and don't get told that it could be important for their future career, no matter how intelligent the kid is
16
9
276
Introducing "HyperFast: Instant Classification for Tabular Data" at @RealAAAI, which received the Best Paper Award at @NeurIPSConf Table rep. workshop @TrlWorkshop! We provide easy-to-use sklearn-like code: https://t.co/qrMF6XStAA Some insights of the work below 👇🧵(1/N)
github.com
HyperFast : Instant Classification for Tabular Data - AI-sandbox/HyperFast
1
3
14
This work has been led by @d_bonet with the supervision of @DocXavi and @alexGioannidis! Code available at: https://t.co/qrMF6XStAA
#AI #AAAI #AAAI2024 #NeurIPS2023 #NeurIPS (5/5)
0
2
4
Hyperfast provides competitive results in several tabular classification datasets, even matching boosting-tree-based accuracies! While still far from solving tabular data classification, we believe Hyperfast provides a step forward in NN-based tabular applications! (4/N)
1
2
5
Hyperfast provides multiple mechanisms to scale to both large and high-dimensional datasets and can be easily applied to real-world applications! (3/N)
1
1
4
Hyperfast replaces the slow process of training MLPs with gradient-based methods (e.g. Adam) with a fast hypernetwork that directly predicts the weights of the MLP. The generated MLP typically matches (or even surpasses) the accuracy of those trained with gradient descent. (2/N)
1
1
6
This work has been led by @d_bonet with the supervision of @DocXavi and @alexGioannidis! Code available at: https://t.co/qrMF6XStAA
#AI #AAAI #AAAI2024 #NeurIPS2023 #NeurIPS (5/5)
0
2
4
Hyperfast provides competitive results in several tabular classification datasets, even matching boosting-tree-based accuracies! While still far from solving tabular data classification, we believe Hyperfast provides a step forward in NN-based tabular applications! (4/N)
1
2
5
Hyperfast provides multiple mechanisms to scale to both large and high-dimensional datasets and can be easily applied to real-world applications! (3/N)
1
1
4