´ë·®±¸¸ÅȨ >
ÄÄÇ»ÅÍ/ÀÎÅͳÝ
>
IT Àü¹®¼­
>
ÄÄÇ»ÅÍ°øÇÐ
>
¾Ë°í¸®Áò/ÀڷᱸÁ¶

ÆîÃ帱â
°­È­ÇнÀ ÀÌ·Ð&½Ç½À : ±âÃÊ ¼öÇкÎÅÍ °­È­ÇнÀ ¾Ë°í¸®Áò±îÁö
Á¤°¡ 30,000¿ø
ÆǸŰ¡ 27,000¿ø (10% , 3,000¿ø)
I-Æ÷ÀÎÆ® 1,500P Àû¸³(6%)
ÆǸŻóÅ ÆǸÅÁß
ºÐ·ù ¾Ë°í¸®Áò/ÀڷᱸÁ¶
ÀúÀÚ È²Çö¼®
ÃâÆÇ»ç/¹ßÇàÀÏ ºñÁ¦ÀÌÆÛºí¸¯ / 2023.10.04
ÆäÀÌÁö ¼ö 392 page
ISBN 9791165922450
»óÇ°ÄÚµå 356802898
°¡¿ëÀç°í Àç°íºÎÁ·À¸·Î ÃâÆÇ»ç ¹ßÁÖ ¿¹Á¤ÀÔ´Ï´Ù.
 
ÁÖ¹®¼ö·® :
´ë·®±¸¸Å Àü¹® ÀÎÅÍÆÄÅ© ´ë·®ÁÖ¹® ½Ã½ºÅÛÀ» ÀÌ¿ëÇÏ½Ã¸é °ßÀû¿¡¼­ºÎÅÍ ÇàÁ¤¼­·ù±îÁö Æí¸®ÇÏ°Ô ¼­ºñ½º¸¦ ¹ÞÀ¸½Ç ¼ö ÀÖ½À´Ï´Ù.
µµ¼­¸¦ °ßÀûÇÔ¿¡ ´ãÀ¸½Ã°í ½Ç½Ã°£ °ßÀûÀ» ¹ÞÀ¸½Ã¸é ±â´Ù¸®½Ç ÇÊ¿ä¾øÀÌ ÇÒÀιÞÀ¸½Ç ¼ö ÀÖ´Â °¡°ÝÀ» È®ÀÎÇÏ½Ç ¼ö ÀÖ½À´Ï´Ù.
¸ÅÁÖ ¹ß¼ÛÇØ µå¸®´Â ÀÎÅÍÆÄÅ©ÀÇ ½Å°£¾È³» Á¤º¸¸¦ ¹Þ¾Æº¸½Ã¸é »óÇ°ÀÇ ¼±Á¤À» ´õ¿í Æí¸®ÇÏ°Ô ÇÏ½Ç ¼ö ÀÖ½À´Ï´Ù.

 ´ë·®±¸¸ÅȨ  > ÄÄÇ»ÅÍ/ÀÎÅÍ³Ý  > IT Àü¹®¼­  > ÄÄÇ»ÅÍ°øÇÐ  > ¾Ë°í¸®Áò/ÀڷᱸÁ¶

 
Ã¥³»¿ë
¡°µ¥ÀÌÅÍ ºÐ¼®¿¡ ÀÇÁ¸ÇÑ ¸Ó½Å·¯´×Àº ÀÌÁ¦ ±×¸¸!¡± ¡º°­È­ÇнÀ ÀÌ·Ð&½Ç½À¡»ÀÌ Ã¥Àº ¾ÆÁ÷µµ ÀΰøÁö´ÉÀÌ Àû¿ëµÇÁö ¸ø ÇÑ »ê¾÷¿¡¼­ ±â¼úÀ» Àû¿ëÇÒ °¡´É¼ºÀ» Á¦¾ÈÇÏ´Â µµ¼­ÀÌ´Ù. ÀÌ Ã¥À̸é Àΰ£º¸´Ù ÁÖ¾îÁø ȯ°æ¿¡ ´õ Àß ÀûÀÀÇÏ´Â ÀΰøÁö´ÉÀÇ °³³äÀ» °øºÎÇÏ¿© ³ª¸¸ÀÇ ÀΰøÁö´ÉÀ» ¸¸µå´Â ±â¼ú¿¡ ´ëÇؼ­ ¹è¿ï ¼ö ÀÖ´Ù.
¸ñÂ÷
ÇÁ·Ñ·Î±× 1. ȯ°æ ¼³Á¤ 1.1 À©µµ¿ì ¹öÀü 1.2 ¸®´ª½º ¹öÀü 2. °­È­ ÇнÀÀ» À§ÇÑ »çÀü Áö½Ä 2.1 ¸Ó½Å ·¯´×°ú °­È­ ÇнÀ 2.1.1 ¸Ó½Å ·¯´× 2.1.2 °­È­ ÇнÀ 2.2 ±âÃÊ ¼öÇÐ 2.2.1 ±âÃÊ ¼±Çü ´ë¼ö 2.2.2 ±âÃÊ ¹ÌºÐ°ú ÀûºÐ 2.2.3 ±âÃÊ È®·ü Åë°è 2.3 ÃÖÀûÈ­ 2.3.1 ´ºÅÏ-·¦½¼¹ý(Newton-Raphson method) 2.3.2 °æ»ç ÇÏ°­¹ý(Gradient descent method) 2.4 ¸ñÀû ÇÔ¼ö 2.4.1 ÃÖ¼Ò Á¦°ö 2.4.2 È®·ü ¿£Æ®·ÎÇÇ¿Í Äð¹é-¶óÀÌºí·¯ ¹ß»ê 2.5 Àΰø ½Å°æ¸Á 2.5.1 ½ÅÈ£ Àü¡¤ÈÄ Ã³¸® 2.5.2 ¼ø¹æÇâ ÀüÆÄ 2.5.3 ¿ª¹æÇâ ÀüÆÄ 2.6 ÃÊ°£´Ü ÆÄÀÌÅäÄ¡ Æ©Å丮¾ó 2.6.1 MNIST 2.6.2 ȸ±Í ºÐ¼® 2.7 ¸Å°³ º¯¼ö Ž»ö¹ý 2.7.1 °ÝÀÚ Å½»ö¹ý(Grid search) 2.7.2 º£ÀÌÁö¾È Ž»ö¹ý(Bayesian optimization) 3. ¸¶¸£ÄÚÇÁ ÀÇ»ç °áÁ¤°ú µ¿Àû °èȹ¹ý Ç®ÀÌ Àü·« 3.1 ¸¶¸£ÄÚÇÁ ÀÇ»ç °áÁ¤ 3.2 µ¿Àû °èȹ¹ý 3.3 [½Ç½À] ÀèÀÇ ·»ÅÍÄ« ¾÷ü ¿î¿µ Àü·« - µ¿Àû °èȹ¹ýÀ» ÀÌ¿ëÇÑ ¸¶¸£ÄÚÇÁ ÀÇ»ç °áÁ¤ 4. ¹ë¸¸ ¹æÁ¤½ÄºÎÅÍ °­È­ ÇнÀ±îÁö 4.1 ¸óÅ×-Ä«¸¦·Î ÃßÁ¤¹ý 4.2 ½Ã°£Â÷ ÇнÀ 4.2.1 TD(0) 4.2.2 TD(¥ë) 4.3 Monte-Carlo vs Temporal Difference 4.4 ¿¡ÀÌÀüÆ® ÇнÀ 4.4.1 SARSA 4.4.2 Q-learning 4.4.3 ½Ç½À 5. Q-ÇÔ¼ö´Â ½Å°æ¸Á¿¡ ¸Ã±ä´Ù - DQN 5.1 DQN 208 5.1.1 ÀÌ·Ð 209 5.1.2 ½Ç½À 219 5.2 ÆÄ»ý ¾Ë°í¸®Áò 256 5.2.1 DDQN 256 5.2.2 PER 260 6. Áï°¢ÀûÀÎ ÇнÀÀÌ ÇÊ¿äÇÒ ¶§ - Policy gradient 6.1 Actor-Critic 6.1.1 ÀÌ·Ð 6.1.2 ½Ç½À 6.2 ÆÄ»ý ¾Ë°í¸®Áò 6.2.1 Asynchronous Advantage Actor-Critic 6.2.2 LSTM-Based Advantage Actor-Critic 6.2.3 [°í±Þ] Trust Region Policy Optimization 6.2.4 [°í±Þ] Proximal Policy Optimization 7. ŽÇèÀÇ Àü·« - Model based learning 7.1 »çÀü Áö½Ä - ¹êµ÷ ¸ðµ¨ 7.2 ÀÌ·Ð - Monte-Carlo Tree Search 7.3 ½Ç½À 7.3.1 CartPole 7.3.2 Tic-Tac-Toe ã¾Æº¸±â

ÀúÀÚ
ȲÇö¼®
ºÎ»ê´ëÇб³¿¡¼­ ÀÚ¿¬°úÇÐÀ» °øºÎÇÏ°í ¼­¿ï´ëÇб³¿¡¼­ °è»ê°úÇÐ ºÐ¾ß·Î Àü°øÀ» ¹Ù²Ù¾î °øÇм®»ç ÇÐÀ§¸¦ ÃëµæÇß´Ù. °ø½Ä ±³°ú°úÁ¤¿¡¼­ Æí¹ÌºÐ ¹æÁ¤½ÄÀ» Ǫ´Â ¹æ½ÄÀ» ¹è¿ï ¶§ ´ëÇпø ½ÃÀý µ¿¾È ¿·ÀÚ¸®¿¡ ¾É¾Ò´ø µ¿±â°¡ ÀΰøÁö´ÉÀ» °øºÎÇÏ´Â °ÍÀ» º¸°í µû¶ó¼­ ¸Ó½Å·¯´×À» Á¶±Ý¾¿ µ¶ÇÐÇϱ⠽ÃÀÛÇß´Ù. ¸Ó½Å·¯´× ºÐ¾ß Áß °ÔÀÓÀ» ½º½º·Î Ç÷¹ÀÌ ÇÏ´Â ÀΰøÁö´ÉÀ» º¸°í °­È­ÇнÀ¿¡ °¡Àå Å« Èï¹Ì¸¦ ´À³¤ µÚ Ãë¹Ì »ï¾Æ ÀÌ·ÐÀ» °øºÎÇÏ°í °³³äÀ» ±¸ÇöÇÏ´Â µ¥ ½Ã°£À» ¸¹ÀÌ º¸³Â´Ù. ¾î¼´Ù º¸´Ï ÇöÀç ÀΰøÁö´É ¾Ë°í¸®ÁòÀ» Àû¿ëÇÏ´Â ¹ÝµµÃ¼ ¸ðµ¨¸µÀ» Ÿ°ÙÀ¸·Î Çϴ ȸ»ç¿¡¼­ ÀΰøÁö´ÉÀ» ÀüÀÚȸ·Î ½Ã¹Ä·¹ÀÌÅÍ¿¡ Á¢¸ñÇÏ´Â ¾÷¹«¸¦ ¸Ã°í ÀÖ´Ù.

ÀÌ ÃâÆÇ»çÀÇ °ü·Ã»óÇ°
°¡Àå ºü¸¥ Ç®½ºÅÃÀ» À§ÇÑ Flask & FastAPI | Dave Lee | ºñÁ¦ÀÌÆÛºí¸¯
ÁÙ¸®¾Æ ÇÁ·Î±×·¡¹Ö | Àü±âÇö,·ù´ë½Ä | ºñÁ¦ÀÌÆÛºí¸¯
ÄÜÅÙÃ÷°¡ µ¸º¸ÀÌ´Â SNS µðÀÚÀÎ with ¹Ì¸®Äµ¹ö½º | ±èÀ¯¼± | ºñÁ¦ÀÌÆÛºí¸¯
âÀÛÀÚ¸¦ À§ÇÑ ¿É½Ãµð¾ð ¸¶½ºÅÍºÏ | Á¶ÁöÈÆ | ºñÁ¦ÀÌÆÛºí¸¯
³ë¼Ç ´öÈÄÀÇ ³î¶ó¿î ²ÜÆÁ ¾ÆÄ«À̺ê | Àü¼Ò¶ó | ºñÁ¦ÀÌÆÛºí¸¯

ÀÌ ºÐ¾ß ½Å°£ °ü·Ã»óÇ°
±×¸²À¸·Î ÀÌÇØÇÏ´Â ¾Ë°í¸®Áò | ¹Ì¾ßÀÚÅ° ½´ÀÌÄ¡,À̽ô٠¸ð¸®Å×·ç,À̵¿±Ô | ±æ¹þ
 
µµ¼­¸¦ ±¸ÀÔÇϽŠ°í°´ ¿©·¯ºÐµéÀÇ ¼­ÆòÀÔ´Ï´Ù.
ÀÚÀ¯·Î¿î ÀÇ°ß ±³È¯ÀÌ °¡´ÉÇÕ´Ï´Ù¸¸, ¼­ÆòÀÇ ¼º°Ý¿¡ ¸ÂÁö ¾Ê´Â ±ÛÀº »èÁ¦µÉ ¼ö ÀÖ½À´Ï´Ù.

µî·ÏµÈ ¼­ÆòÁß ºÐ¾ß¿Í »ó°ü¾øÀÌ ¸ÅÁÖ ¸ñ¿äÀÏ 5ÆíÀÇ ¿ì¼öÀÛÀ» ¼±Á¤ÇÏ¿©, S-Money 3¸¸¿øÀ» Àû¸³Çص帳´Ï´Ù.
ÃÑ 0°³ÀÇ ¼­ÆòÀÌ ÀÖ½À´Ï´Ù.