DataTalks #13: The Beauty and The Hyperparameter Optimization Beast
Details
Our 13th meetup is hosted by 𝗪𝗶𝘅 and it's going to be the most awesome one yet!
𝗪𝗶𝘅 is going to share with us a never before publicly shared journey of discovering an optimal solution for visual beauty evaluation.
Next, Gilad Rubin will blow our mind with a super practical talk about a new and innovative package to tune hyperparameters 🤯
🆒🆒🆒
𝘼𝙜𝙚𝙣𝙙𝙖:
🍕 18:30 - 19:00 - Gathering, snacks & mingling
💎 19:00 - 19:45 - First talk:
Can Machines Learn Beauty?
by Olha Shainoha
(English)
☄️ 19:55 - 20:40 - Second talk:
HyPSTER - HyperParameter optimization on STERoids
by Gilad Rubin
(English)
𝘾𝙖𝙣 𝙈𝙖𝙘𝙝𝙞𝙣𝙚𝙨 𝙇𝙚𝙖𝙧𝙣 𝘽𝙚𝙖𝙪𝙩𝙮? - Olha Shainoha (English)
How can we objectify beauty? Asking whether an image is beautiful, or how beautiful it is, will probably revoke different responses from different individuals. “Beauty is in the eye of the beholder”, it’s a subjective thing. So, is it possible to teach a machine to make this decision if we humans fail to agree? My answer is - yes. In this talk, I’ll share our journey of discovering an optimal solution for visual beauty evaluation. The transition from Domain experts to Crowdsourcing, from Likert scale to Pairwise comparison, from Ranking to Binary labels, and introduce the final framework we developed based on the acquired knowledge.
𝙃𝙮𝙋𝙎𝙏𝙀𝙍 - 𝙃𝙮𝙥𝙚𝙧𝙋𝙖𝙧𝙖𝙢𝙚𝙩𝙚𝙧 𝙤𝙥𝙩𝙞𝙢𝙞𝙯𝙖𝙩𝙞𝙤𝙣 𝙤𝙣 𝙎𝙏𝙀𝙍𝙤𝙞𝙙𝙨 - Gilad Rubin (English)
You're a good Data Scientist. You know that you need to work on your data - gathering, exploring, cleaning, processing & validating it.
Often times, the best way to evaluate your work is by constructing an ML Pipeline and checking how it performs on various metrics.
Since Data Science is in its nature an experimental and iterative process - we need to ask ourselves: How can we choose a proper ML Pipeline efficiently to allow for rapid and continuous improvement on our projects?
What if I told you there was a way to find a simple and accurate ML Pipeline using just a few lines of code with blazing fast performance?
Well, dream no more.
Introducing HyPSTER - HyperParameter Optimization on STERoids
HyPSTER is a brand new Python package built on top of Optuna (an awesome Hyperparameter Optimization framework) that helps you find compact and accurate ML Pipelines while staying light and efficient.
HyPSTER uses state of the art algorithms for sampling hyperparameters (e.g. TPE, CMA-ES) and pruning unpromising trials (e.g. Asynchronous Successive Halving), combined with cross-validated early stopping and adaptive learning rates, all packed up in a simple sklearn API that allows for automatic Preprocessing pipeline selection and supports your favorite ML packages (e.g. XGBoost, LightGBM, CatBoost, SGDClassifier) out of the box. And yes, it supports multi CPU/GPU training.
In this talk, I will try to explain the internals of the package and demonstrate its power and simplicity using real-world datasets.
