ํ‹ฐ์Šคํ† ๋ฆฌ ๋ทฐ

ใ…‡ ๊ต์ˆ˜ : ์ดํ™”์—ฌ์ž๋Œ€ํ•™๊ต ๊ฐ•์ œ์› ๊ต์ˆ˜ 

ใ…‡ ํ•™์Šต๋ชฉํ‘œ 

     Machine Learning์˜ ํ•œ ๋ถ€๋ฅ˜์ธ ์ง€๋„ํ•™์Šต(Supervised Learning)์— ๋Œ€ํ•œ ๊ธฐ๋ณธ ๊ฐœ๋…๊ณผ regression/classification์˜ ๋ชฉ์  ๋ฐ ์ฐจ์ด์ ์— ๋Œ€ํ•ด ์ดํ•ดํ•˜๊ณ , 

     ๋‹ค์–‘ํ•œ ๋ชจ๋ธ ๋ฐ ๋ฐฉ๋ฒ•๋ก (linear and nonlinear regression, classification, ensemble methods, kernel methods ๋“ฑ)์„ ํ†ตํ•ด 

     ์–ธ์ œ ์–ด๋–ค ๋ชจ๋ธ์„ ์‚ฌ์šฉํ•ด์•ผ ํ•˜๋Š”์ง€, ์™œ ์‚ฌ์šฉํ•˜๋Š”์ง€, ๋ชจ๋ธ ์„ฑ๋Šฅ์„ ํ–ฅ์ƒ์‹œํ‚ค๋Š” ๋ฐฉ๋ฒ•์„ ํ•™์Šตํ•˜๊ฒŒ ๋ฉ๋‹ˆ๋‹ค. 

 


Part 1. SL Foundation

- ์ง€๋„ํ•™์Šต : ๋ถ„๋ฅ˜ / ํšŒ๊ท€ ๊ฐ€ ์žˆ์Œ

- error๋ฅผ ๊ณ„์‚ฐํ•ด์„œ ์†์‹คํ•จ์ˆ˜ loss function cost function์œผ๋กœ ๋ถ€๋ฆ„

- squared error, binary error

- Curse of dimension ์ฐจ์›์˜ ์ €์ฃผ ๋ฌธ์ œ (์˜ค๋ฒ„ํ”ผํŒ…)

- ๋ฐ์ดํ„ฐ๋ฅผ ๋Š˜๋ฆฌ๋Š” ๊ฒƒ์ด ๊ฐ€์žฅ ๋‹จ์ˆœํ•œ ํ•ด๊ฒฐ ๋ฐฉ๋ฒ•

- data augmentation! ์—ฌ๋Ÿฌ ๋ฐฉ๋ฒ• : k-fold cross validation

- Regularization ๋ฐฉ๋ฒ•

- ์•™์ƒ๋ธ” ๋ฐฉ๋ฒ•

 

Part 2. Linear Regression

- Linear models

- Least square problem (๋ฐฉ์ •์‹ : Normal equation) -> ์ตœ์ ์˜ ์„ธํƒ€๋ฅผ ๊ตฌํ•˜๊ฒŒ ๋จ.

- Gradient : ํ•จ์ˆ˜๋ฅผ ๋ฏธ๋ถ„ํ•˜์—ฌ ์–ป๋Š” term์œผ๋กœ ํ•ด๋‹น ํ•จ์ˆ˜์˜ ๋ณ€ํ™”ํ•˜๋Š” ์ •๋„๋ฅผ ํ‘œํ˜„ํ•˜๋Š” ๊ฐ’

- Gradient descent : ํ•จ์ˆ˜์˜ ๋ณ€ํ™”๋„๊ฐ€ ๊ฐ€์žฅ ํฐ ๋ฐฉํ–ฅ์œผ๋กœ ์ด๋™

 

Part 3. Gradient descent

- SGD : noise์˜ํ–ฅ ๋ฐ›๊ธฐ ์‰ฌ์›€

- Local Optimum ํ•œ๊ณ„

 

Part 4. Linear classification

- ๋ถ„๋ฅ˜๋ฌธ์ œ

- data set์•ˆ์˜ positive sample๊ณผ negative sample

- Zero - one loss

- ๋ฌธ์ œ์  ํ•ด๊ฒฐ ์œ„ํ•ด hinge loss ์‚ฌ์šฉ

- Cross-entropy loss

 

Part 5. Advanced Classification

- SVM : Margin ๊ฐœ๋… ์ค‘์š” support vector -> ์„ฑ๋Šฅ์„ ์ขŒ์ง€์šฐ์ง€

- Optimization : Hard Margin SVM

- Kernel ํ•จ์ˆ˜ : ์ฐจ์ˆ˜๋ฅผ ๋†’์—ฌ linearํ•˜๊ฒŒ ๋งŒ๋“œ๋Š”

- ANN

- Back Propagation : dropout

 

Part 6. Ensemble

- ๋จธ์‹ ๋Ÿฌ๋‹์—์„œ ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ ์ข…๋ฅ˜์— ์ƒ๊ด€์—†์ด ์„œ๋กœ ๋‹ค๋ฅด๊ฑฐ๋‚˜, ๊ฐ™์€ ๋งค์ปค๋‹ˆ์ฆ˜์œผ๋กœ ๋™์ž‘ํ•˜๋Š” ๋‹ค์–‘ํ•œ ๋จธ์‹ ๋Ÿฌ๋‹ ๋ชจ๋ธ์„ ๋ฌถ์–ด ํ•จ๊ป˜ ์‚ฌ์šฉํ•˜๋Š” ๋ฐฉ์‹

- ๋‹ค์–‘ํ•œ ๋ชจ๋ธ์˜ ๊ฐ ์žฅ์ ์„ ์‚ด๋ ค ์˜ˆ์ธก ์„ฑ๋Šฅ์„ ํ–ฅ์ƒํ•  ์ˆ˜ ์žˆ์Œ

- ๊ฐ€์žฅ ๊ธฐ๋ณธ์ ์ธ ์š”์†Œ ๊ธฐ๋Šฅ์€ bagging ๊ณผ boosting

- ์žฅ์  : ์˜ˆ์ธก ์„ฑ๋Šฅ์„ ์•ˆ์ •์ ์œผ๋กœ ํ–ฅ์ƒ ๊ฐ€๋Šฅ, ์‰ฝ๊ฒŒ ๊ตฌํ˜„์ด ๊ฐ€๋Šฅ

- bagging : ํ•™์Šต ๊ณผ์ •์—์„œ training data๋ฅผ ๋žœ๋คํ•˜๊ฒŒ ๋‚˜๋ˆ ์„œ ํ•™์Šต -> "๋ชจ๋ธ์„ ๋ณ‘๋ ฌ์ ์œผ๋กœ ํ•™์Šต ๊ฐ€๋Šฅ" (bootstrapping / aggregating)

- boosting : ์‹œํ€€์…œํ•˜๊ฒŒ ๋™์ž‘ํ•จ, classfier์„ ์—ฐ์†์ ์œผ๋กœ ๋™์ž‘ํ•˜๊ฒŒ ํ•จ (cascading of weak classifier)

- ๋Œ€ํ‘œ์ ์œผ๋กœ Adaboost ๊ฐ€ ์žˆ์Œ. (์˜ค๋ถ„๋ฅ˜๋œ sample์— ๋†’์€ ๊ฐ€์ค‘์น˜๋ฅผ ์ค˜์„œ ๋‹ค์Œ ํ•™์Šต์— ์‚ฌ์šฉํ•  ์ˆ˜ ์žˆ๊ฒŒํ•จ -> ๋‹ค์Œ classifier๋Š” ์–ด๋ ค์šด ์ƒ˜ํ”Œ์„ ํ•ด๊ฒฐํ•˜๋Š”๋ฐ ๋” ํŠนํ™”๋œ ๋ชจ๋ธ๋กœ ๋™์ž‘ํ•˜๊ฒŒ ๋จ.)

- Random forest (๋ฐฐ๊น…์„ ํ†ตํ•ด ํ•™์Šต, ๋ถ€์ŠคํŒ… ์‚ฌ์šฉ)

- Evaluation : Accuracy์ธก์ •, confusion matrix (์˜ค์ฐจ๊ณ„์‚ฐ), ROC Curve

 


์žฌ์ž‘๋…„ ๊ธฐ๊ณ„ํ•™์Šต ์ˆ˜์—…์‹œ๊ฐ„์— ๋ฐฐ์› ๋˜ ์ด๋ก ๋“ค์ด ๋– ์˜ค๋ฅธ๋‹ค.. ๊ทธ๋•Œ ์—ด์‹ฌํžˆ ๊ณต๋ถ€ํ–ˆ์—ˆ๋Š”๋ฐ ์‚ฌ์‹ค ๊ธฐ์–ต์ด ์ž˜ ์•ˆ๋‚จ

๊ทธ๋ž˜๋„ ์ต์ˆ™ํ•œ ๋ชจ๋ธ๋“ค์ด๋ž‘ ๋ฐฉ๋ฒ•๋ก ์ด ๋ณด์ด๋Š” ๊ฑธ ๋ณด๋‹ˆ๊ฐ€ ์œ ๊ฒฝ์ดˆ์ด์˜ ์ˆ˜์—…์€ ์ •๋ง ๋งŽ์€ ๋„์›€์ด ๋์—ˆ๊ตฌ๋‚˜ ์‹ถ์Œ.

์ธ๊ณต์ง€๋Šฅ์„ ์ข€ ๋” ์—ด์‹ฌํžˆ ๋“ค์„๊ฑธ! ํ•˜๋Š” ํ›„ํšŒ๊ฐ€ ๋‚จ์ง€๋งŒ ๊ทธ๋• ๋„ˆ๋ฌด ๋ฐ”๋นณ๋‹ค.....ใ„ฑ-

๋ฐ˜์‘ํ˜•

'AI > LG Aimers' ์นดํ…Œ๊ณ ๋ฆฌ์˜ ๋‹ค๋ฅธ ๊ธ€

Module 5. ์ธ๊ณผ์ถ”๋ก   (0) 2024.01.14
Module 3. Machine Learning ๊ฐœ๋ก   (2) 2024.01.09
Module 2. Mathematics for ML  (0) 2024.01.09
Module 1. AI ์œค๋ฆฌ  (0) 2024.01.05