Explore tweets tagged as #hyperparameter
@davisblalock
Davis Blalock
20 days
Deep learning training is a mathematical dumpster fire. But it turns out that if you *fix* the math, everything kinda just works…fp8 training, hyperparameter transfer, training stability, and more. [1/n]
Tweet media one
14
150
1K
@InformationMDPI
Information MDPI
5 hours
Read #HighlyAccessedArticle "Structure Learning and Hyperparameter Optimization Using an Automated Machine Learning (AutoML) Pipeline". See more details at: .#Bayesianoptimization .#hyperparameteroptimization.@ComSciMath_Mdpi
Tweet media one
0
3
4
@UncertaintyInAI
UAI 2025
3 hours
and . "Hyperparameter Optimization and Algorithm Selection: Practical Techniques, Theory, and New Frontiers". by Dravyansh Sharma (TTIC)
Tweet media one
Tweet media two
@UncertaintyInAI
UAI 2025
3 hours
time to educate our crowds on . "Counterfactuals in Minds and Machines". by @tobigerstenberg (Stanford), @autreche (MPI SS) and @stratis_ (MPI SS)
Tweet media one
Tweet media two
0
3
4
@gm8xx8
𝚐𝔪𝟾𝚡𝚡𝟾
1 month
Don't be lazy: CompleteP enables compute-efficient deep transformers. - Sets α = 1 (no residual scaling) and rescales LR, init, LayerNorm, bias, ε, and weight decay to retain training stability.- Enables true hyperparameter transfer across width and depth.- Prevents layers from
Tweet media one
3
27
158
@EconoMind__
Morales
26 days
📢 Save Time on Hyperparameter Tuning with Hyperband!. Hyperparameter search can easily eat up days of compute and leave you drowning in experiments. Enter Hyperband, a bandit-based early-stopping strategy that lets you:. 1️⃣ What It Is. A smart scheduler that tests many
Tweet media one
0
1
12
@TheoW0lf
Theo Wolf
18 days
🚀 Excited to announce Hyperoptax, a library for parallel hyperparameter tuning in JAX. Implements Grid, Random, and Bayesian search in pure JAX so that you can rapidly search across parameter configurations in parallel ‖. 📦 pip install hyperoptax.
3
6
48
@arjunkocher
Arjun
28 days
Muon optimizer doesn’t consistently outpace AdamW in grokking. Finding by:. The expanded hyperparameter sweep, tweaking transformer embedding dimensions and batch sizes reveals that validation accuracy curves for both optimizers converge similarly after
Tweet media one
1
2
19
@Soledad_Galli
Soledad Galli
6 days
Here are some of the most widely used techniques for hyperparameter optimization. The table below highlights the pros and cons of each strategy, helping you better understand when to use each one. Hope you find it helpful! 😉 . #machinelearning #datascience #ML #MLModels
Tweet media one
1
0
11
@Maximize_AI
Maximize AI
11 days
2 / MaxiScreen Dashboard (Phase 1). - Live Bittensor subnet metagraph visualization.- Deep hyperparameter insights (tempo scores, emission rates, and more).- Intuitive interface designed for everyone.- Technical complexity made accessible to non-technical users. Beta release
Tweet media one
11
14
25
@JoachimSchork
Joachim Schork
13 days
Bayesian optimization is a method used to optimize complex functions that are expensive or time-consuming to evaluate. It is widely applied in machine learning, engineering, and scientific research to improve efficiency and precision, particularly in scenarios like hyperparameter
0
25
180
@c_voelcker
Claas Voelcker
3 days
The result is REPPO, which trains as fast as PPO, without replay buffers, and with minimal hyperparameter tuning. If you don't believe us, take our code and test it! We provide implementations in both jax and torch (but jax is faster 😜):
Tweet media one
Tweet media two
1
0
6
@MathematicsMDPI
Mathematics MDPI
25 days
🎉WoS #HighlyCited.🔖TPTM-HANN-GA: A Novel #Hyperparameter_Optimization Framework Integrating the #Taguchi_Method, an #Artificial_Neural_Network, and a #Genetic_Algorithm for the Precise Prediction of #Cardiovascular_Disease_Risk.👥by Chia-Ming Lin et al.🔗
Tweet media one
0
0
0
@Soledad_Galli
Soledad Galli
11 days
🤔 Not sure which hyperparameter search method to use? . - Random Search.- Bayesian Search.- SMAC.- TPE (Tree-structured Parzen Estimator). Watch the video for a quick rundown 👇. #machinelearning #smac #mlmodels #hyperparameter #tpe #randomsearch  #bayesiansearch
0
0
1
@Ava_AITECH
AvaChat
11 days
Unlocking AI Potential with Python 🚀. Discover the power of PerpetualBooster, a gradient boosting algorithm that doesn't need hyperparameter tuning! 🤖💻 What's your favorite AI Python library?.
0
23
82
@Soledad_Galli
Soledad Galli
26 days
Grid Search 🆚 Random Search: Two powerful methods for hyperparameter tuning in Machine Learning. Here's a chart for a side-by-side comparison of their pros and cons. #DataScience #AI #ML #Machinelearning #hyperparameter #gridsearch #randomsearch
Tweet media one
0
0
3
@Rahuladi11
Aditya | 𝔽rAI
11 days
Picture this: You're at the starting line, heart racing—and Hyperbolic Labs whispers, “Hyperparameter tuning isn’t just for the pros. It’s for you.” 🌟. Imagine optimizing AI like you're fine-tuning your own playlist—smooth, personal, powerful. No PhD? No problem.
Tweet media one
1
0
2
@mariofilhoml
Mario Filho
25 days
This has become my favorite hyperparameter in GBDT-based models. Borrows fundamental statistical thinking to combat overfitting
Tweet media one
0
0
3
@arjunverma2004
Arjun Verma
3 days
Day 48 of #100DaysOfCode .-> Today I learnt about Hyperparameter tuning in Keras.-> Then practiced it on PIMA Indians diabetes dataset.-> Got accuracy of 0.77 although I didn't preprocessed the data
Tweet media one
Tweet media two
0
0
2