ML Intro

Machine Learning

  • ์ฃผ์–ด์ง„ ๋ฐ์ดํ„ฐ๋ฅผ ํ†ตํ•ด์„œ ์ž…๋ ฅ๋ณ€์ˆ˜์™€ ์ถœ๋ ฅ๋ณ€์ˆ˜ ๊ฐ„์˜ ๊ด€๊ณ„๋ฅผ ๋งŒ๋“œ๋Š” ํ•จ์ˆ˜ $f$๋ฅผ ๋งŒ๋“œ๋Š” ๊ฒƒ
  • ์ฃผ์–ด์ง„ ๋ฐ์ดํ„ฐ ์†์—์„œ ๋ฐ์ดํ„ฐ์˜ ํŠน์ง•์„ ์ฐพ์•„๋‚ด๋Š” ํ•จ์ˆ˜ $f$๋ฅผ ๋งŒ๋“œ๋Š” ๊ฒƒ

1. ๊ธฐ๋ณธ ๊ฐœ๋…๊ตฌ๋ถ„

  • ์ง€๋„ ํ•™์Šต: ํšŒ๊ท€(Regression), ๋ถ„๋ฅ˜(Classification)
  • ๋น„์ง€๋„ ํ•™์Šต: PCA, ๊ตฐ์ง‘๋ถ„์„
  • ๊ฐ•ํ™” ํ•™์Šต: ์ˆ˜๋งŽ์€ ์‹œ๋ฎฌ๋ ˆ์ด์…˜์„ ํ†ตํ•ด ํ˜„์žฌ์˜ ์„ ํƒ์ด ๋จผ ๋ฏธ๋ž˜์— ๋ณด์ƒ์ด ์ตœ๋Œ€๋กœ ํ•˜๋Š” action์„ ํ•™์Šต

2. ๋‹ค์–‘ํ•œ ๋จธ์‹ ๋Ÿฌ๋‹ ๊ธฐ๋ฒ•

  1. ์„ ํ˜•ํšŒ๊ท€๋ถ„์„: ์„ ํ˜•๊ด€๊ณ„๋ฅผ ๊ฐ€์ •ํ•˜์—ฌ, ๋…๋ฆฝ๋ณ€์ˆ˜์˜ ์ค‘์š”๋„์™€ ์˜ํ–ฅ๋ ฅ ํŒŒ์•…
  2. DT(Decision Tree): ๋…๋ฆฝ๋ณ€์ˆ˜์˜ ์กฐ๊ฑด์— ๋”ฐ๋ผ ์ข…์†๋ณ€์ˆ˜๋ฅผ ๋ถ„๋ฆฌ
  3. KNN(K-Nearest Neighbor): ์ƒˆ๋กœ ๋“ค์–ด์˜จ ๋ฐ์ดํ„ฐ์˜ ์ฃผ๋ณ€ K๊ฐœ์˜ ๋ฐ์ดํ„ฐ์˜ class๋กœ ๋ถ„๋ฅ˜
  4. NN(Neural Network): ์ž…๋ ฅ์ธต/์€๋‹‰์ธต/์ถœ๋ ฅ์ธต ์œผ๋กœ ๊ตฌ์„ฑ๋œ ๋ชจํ˜•. ๊ฐ ์ธต์„ ์—ฐ๊ฒฐํ•˜๋Š” ๋…ธ๋“œ์˜ ๊ฐ€์ค‘์น˜๋ฅผ ์—…๋ฐ์ดํŠธํ•˜๋ฉฐ ํ•™์Šต
  5. SVM(Support Vector Machine): class ๊ฐ„ ๊ฑฐ๋ฆฌ๊ฐ€ ์ตœ๋Œ€๊ฐ€ ๋˜๋„๋ก decision boundary ๋งŒ๋“œ๋Š” ๋ฐฉ๋ฒ•
  6. K-means Clustering: Label ์—†์ด ๋ฐ์ดํ„ฐ์˜ ๊ตฐ์ง‘ k๊ฐœ ์ƒ์„ฑ
  7. Ensemble Learning: ์—ฌ๋Ÿฌ ๊ฐœ์˜ ๋ชจ๋ธ์„ ๊ฒฐํ•ฉํ•˜์—ฌ ์‚ฌ์šฉํ•˜๋Š” ๋ชจ๋ธ๋กœ, ๊ตฌ์ฒด์ ์œผ๋กœ๋Š” ๋‹ค์–‘ํ•œ ์•Œ๊ณ ๋ฆฌ์ฆ˜ ์ข…๋ฅ˜๊ฐ€ ์žˆ๋‹ค.
    7-1. Bagging: ๋ชจ๋ธ์„ ๋‹ค์–‘ํ•˜๊ฒŒ ๋งŒ๋“ค๊ธฐ ์œ„ํ•ด ๋ฐ์ดํ„ฐ๋ฅผ ์žฌ๊ตฌ์„ฑ
    7-2. Random Forest: ๋ชจ๋ธ์„ ๋‹ค์–‘ํ•˜๊ฒŒ ๋งŒ๋“ค๊ธฐ ์œ„ํ•ด ๋ฐ์ดํ„ฐ๋ฟ๋งŒ ์•„๋‹ˆ๋ผ ๋ณ€์ˆ˜๋„ ์žฌ๊ตฌ์„ฑ
    7-3. Boosting: ๋งž์ถ”๊ธฐ ์–ด๋ ค์šด ๋ฐ์ดํ„ฐ์— ๋Œ€ํ•ด ์ข€ ๋” ๊ฐ€์ค‘์น˜๋ฅผ ๋‘์–ด seqeuntialํ•˜๊ฒŒ ํ•™์Šตํ•˜๋Š” ๊ฐœ๋… (ex. AdaBoost, Gradient Boosting(Xgboost, LightGBM, CatBoost)
    7-4. Stacking: ๋ชจ๋ธ์˜ output๊ฐ’์„ ์ƒˆ๋กœ์šด ๋…๋ฆฝ๋ณ€์ˆ˜๋กœ ํ™œ์šฉ
  8. Deep Learning: ๋”ฅ๋Ÿฌ๋‹์€ ์‚ฌ์‹ค ๋จธ์‹ ๋Ÿฌ๋‹์˜ ๋ถ€๋ถ„์ง‘ํ•ฉ์ด๋‹ค. ํ•˜์ง€๋งŒ ์›Œ๋‚™ ๊นŠ๊ณ  ๋‹ค์–‘ํ•˜๊ธฐ์— ๋”ฐ๋กœ ๋‹ค๋ฃจ๋„๋ก ํ•˜๊ฒ ๋‹ค.

3. ๋ชจํ˜•์˜ ์ ํ•ฉ์„ฑ ํ‰๊ฐ€ ๋ฐ ์‹คํ—˜์„ค๊ณ„

๋ฐ์ดํ„ฐ๋ฅผ Training-Validation-Test, ์ด ์„ธ ๊ฐ€์ง€ ์„ธํŠธ๋กœ ๋‚˜๋ˆˆ๋‹ค.

K-Fold Cross Validation

๋ฐ์ดํ„ฐ๋ฅผ k๊ฐœ ๋ถ€๋ถ„์œผ๋กœ ๋‚˜๋ˆ„ ๋’ค, ํ•˜๋‚˜๋ฅผ ๊ฒ€์ฆ์ง‘ํ•ฉ ๋‚˜๋จธ์ง€๋ฅผ ํ•™์Šต์ง‘ํ•ฉ์œผ๋กœ ํ•œ๋‹ค. ์ด ๊ณผ์ •์„ k๋ฒˆ ๋ฐ˜๋ณตํ•ด์„œ k๊ฐœ์˜ ์„ฑ๋Šฅ์ง€ํ‘œ๋ฅผ ๊ตฌํ•˜๊ณ  ๊ทธ๊ฒƒ๋“ค์˜ ํ‰๊ท ์„ ๊ตฌํ•œ๋‹ค.

LOOCV(Leave One Out Cross Validation)

๋ฐ์ดํ„ฐ๋ฅผ k๊ฐœ์˜ ๋ถ€๋ถ„์œผ๋กœ ๋‚˜๋ˆ„๊ธฐ์— ๋ถ€์กฑํ•  ๋•Œ, ๋ฐ์ดํ„ฐ ํ•œ ๊ฐœ์”ฉ์„ ๋นผ๊ฐ€๋ฉด์„œ K-fold CV๋ฅผ ํ•˜๋Š” ๋ฐฉ์‹๊ณผ ๋˜‘๊ฐ™์ด ํ•œ๋‹ค.

4. ๊ณผ์ ํ•ฉ(Overfitting)

๋จธ์‹ ๋Ÿฌ๋‹์—์„œ ๊ฐ€์žฅ ์ฃผ์˜ํ•ด์•ผ ํ•  ๊ฒƒ ์ค‘ ํ•˜๋‚˜๊ฐ€ ๋ฐ”๋กœ ๊ณผ์ ํ•ฉ์ด๋‹ค. ์ด์™€ ๊ด€๋ จํ•ด์„œ๋Š” Bias-Variance Tradeoff์— ๋Œ€ํ•œ ์ดํ•ด๊ฐ€ ํ•„์š”ํ•˜๋‹ค. ๊ทธ๋ฆฌ๊ณ  ์•„์ฃผ ๊ฐ„๋‹จํ•˜๊ฒŒ ์ดํ•ดํ•˜๊ธฐ ์œ„ํ•ด์„œ๋Š” ์•„๋ž˜ ๋‘ ์‚ฌ์ง„์„ ์ฐธ๊ณ ํ•˜๋ฉด ๋  ๊ฒƒ์ด๋‹ค.

overfitting1


overfitting2

์ฐธ๊ณ 

[1] https://medium.com/@cs.sabaribalaji/overfitting-6c1cd9af589
[2] https://www.researchgate.net/figure/The-overfitting-of-model-a-training-error-and-true-error-b-depiction-of-Eq-33_fig5_333505702