캐글

[Kaggle Study] 14. Hyperparameter Tuning

dongsunseng 2024. 11. 16. 23:36
반응형

This post heavily relies on Andrew Ng's lecture: 

 

심층 신경망 개선하기: 하이퍼파라미터 튜닝, 정규화 및 최적화

DeepLearning.AI에서 제공합니다. 딥러닝 스페셜라이제이션의 두 번째 과정에서는 딥러닝 블랙박스를 열어 성능을 향상시키고 좋은 결과를 도출하는 프로세스를 체계적으로 이해합니다. 딥러닝 애

www.coursera.org

Hyperparameter Tuning?

  • The process of adjusting hyperparameters to optimize the model for better performance
  • There are certain hyperparameters that have higer tuning priority than others.
    • In other words, some hyperparameters can work well with commonly known values without tuning, while others require tuning to determine which values work best.
  • There are various tuning methods such as:
    • Manual Search
    • Grid Search
    • Random Search
    • Bayesian Optimization
    • Non-Probabilistic
    • Evolutionary Optimization
    • Gradient-based Optimization
    • Early Stopping
    • and more
  • All methods except Manual Search are called 'Automated Hyperparameter Selection'

  • This image indicates that using random values is much better way than using grid to test hyperparameters.
    • When $\alpha$ is a hyperparameter that requires a lot of tuning, we can only try fixed amount of values if we use grid method. 
    • However, when we use random values, we can try much more variety of values.

  • 'Coarse to fine' search is a process that we narrow down to smaller range(square in the image) when we find several values that work well and then try more densely in that range of values.

Manual Search

  • Also called 'rules of thumb', which referes to setting hyperparameter values based on experience or intuition.
  • In reality, it is convenient to follow commonly known values since they usually perform well and only require minor code adjustments.
  • However, the downside is that it's difficult to compare performance across different hyperparameter combinations.

Grid Search

  •  

 

 

 

 

 

Reference

 

Automated Hyperparameter Tuning

Explore and run machine learning code with Kaggle Notebooks | Using data from multiple data sources

www.kaggle.com

 


Excellence is not a destination; it is a continuous journey that never ends.
- Conor Mcgregor -
반응형