Is this email not displaying correctly? View it in your browser.
Train in Data, learn machine learning online

The Right Method Leads to the Right Library!

Image description

Welcome to Data Bites!



Every Monday, I’ll drop a no-fluff, straight-to-the-point tip on a data science skill, tool, or
method to help you stay sharp in the field. I hope you find it useful!

Build, Deploy, Succeed: Your Weekly Machine Learning Deployment Series

I’m excited to share that I’ve launched a weekly learning playlist on YouTube! 



Each week, a new video will be added so you can follow along at a steady, manageable pace.

Image description

Each video builds on the last, making it easy to stay engaged and make learning part of your weekly routine. Don’t forget to subscribe so you’ll be notified as soon as a new lesson is released.



Thank you for being part of this learning journey. I hope these weekly videos bring clarity, inspiration, and momentum to your practice.

Watch the latest lesson and follow along week by week

Choose the library based on the method and not the method based on the library.

I often hear people say, "I use Hyperopt for hyperparameter tuning," or "Optuna produces the best results."



It almost sounds like these libraries work some kind of magic, but the reality is that they implement well-researched methods. Yes, they’re effective, but they’re not magic—they’re based on math.



Both Optuna and Hyperopt utilize common techniques, including:

  • Bayesian Optimization using Tree Parzen Estimators as surrogates
  • Randomized Search

However, they also have distinct differences:

  • Optuna makes it easier to set up searches
  • Optuna has better documentation
  • Optuna supports additional algorithms, including multifidelity optimization, Bayesian optimization with Gaussian processes, grid search, and genetic algorithms
🧐That said, neither Optuna nor Hyperopt currently supports Bayesian optimization with random forests as a surrogate, which some studies suggest is the best option for hyperparameter tuning.



Additionally, scikit-learn provides randomized search and multifidelity models as well. So why would you use something else when training scikit-learn models?



💡In short, understanding how these libraries work empowers you to choose the one that best fits your needs rather than adapting your needs to fit a library.



Knowledge is power! 💪



🎓Check out my course, Hyperparameter Optimization for Machine Learning, for comprehensive learning.


I hope this information was useful!



Wishing you a successful week ahead - see you next Monday! 👋🏻


Sole

Ready to enhance your skills?

Our specializations, courses and books are here to assist you:

More courses

Did someone share this email with you? Think it's pretty cool? Then just hit the button and subscribe to Data Bites. Don’t miss out on any of our tips and propel your data science career to new heights.

Subscribe
Image description

Hi…I’m Sole



The main instructor at Train in Data. My work as a data scientist, includes creating and implementing machine learning models for evaluating insurance claims, managing credit risk, and detecting fraud. In 2018, I was honoured with a Data Science Leaders' award, and in 2019 and again in 2024, I was acknowledged as one of LinkedIn's voices in data science and analytics.

View

You are receiving this email because you subscribed to our newsletter, signed up on our website, purchased or downloaded any products from us.

Follow us on social media

Copyright (C) 2025 Train in Data. All rights reserved.

If you would like to unsubscribe, please click here.