본문 바로가기

리디 접속이 원활하지 않습니다.
강제 새로 고침(Ctrl + F5)이나 브라우저 캐시 삭제를 진행해주세요.
계속해서 문제가 발생한다면 리디 접속 테스트를 통해 원인을 파악하고 대응 방법을 안내드리겠습니다.
테스트 페이지로 이동하기

Deep Reinforcement Learning Hands-On 상세페이지

컴퓨터/IT 개발/프로그래밍 ,   컴퓨터/IT IT 해외원서

Deep Reinforcement Learning Hands-On

Apply modern RL methods, with deep Q-networks, value iteration, policy gradients, TRPO, AlphaGo Zero and more
소장전자책 정가12,000
판매가12,000
Deep Reinforcement Learning Hands-On 표지 이미지

Deep Reinforcement Learning Hands-On작품 소개

<Deep Reinforcement Learning Hands-On> ▶Book Description
Recent developments in reinforcement learning (RL), combined with deep learning (DL), have seen unprecedented progress made towards training agents to solve complex problems in a human-like way. Google's use of algorithms to play and defeat the well-known Atari arcade games has propelled the field to prominence, and researchers are generating new ideas at a rapid pace.

Deep Reinforcement Learning Hands-On is a comprehensive guide to the very latest DL tools and their limitations. You will evaluate methods including Cross-entropy and policy gradients, before applying them to real-world environments. Take on both the Atari set of virtual games and family favorites such as Connect4. The book provides an introduction to the basics of RL, giving you the know-how to code intelligent learning agents to take on a formidable array of practical tasks. Discover how to implement Q-learning on 'grid world' environments, teach your agent to buy and trade stocks, and find out how natural language models are driving the boom in chatbots.

▶What You Will Learn
⦁ Understand the DL context of RL and implement complex DL models
⦁ Learn the foundation of RL: Markov decision processes
⦁ Evaluate RL methods including Cross-entropy, DQN, Actor-Critic, TRPO, PPO, DDPG, D4PG and others
⦁ Discover how to deal with discrete and continuous action spaces in various environments
⦁ Defeat Atari arcade games using the value iteration method
⦁ Create your own OpenAI Gym environment to train a stock trading agent
⦁ Teach your agent to play Connect4 using AlphaGo Zero
⦁ Explore the very latest deep RL research on topics including AI-driven chatbots

▶Key Features
⦁ Explore deep reinforcement learning (RL), from the first principles to the latest algorithms
⦁ Evaluate high-profile RL methods, including value iteration, deep Q-networks, policy gradients, TRPO, PPO, DDPG, D4PG, evolution strategies and genetic algorithms
⦁ Keep up with the very latest industry developments, including AI-driven chatbots

▶Who This Book Is For
The main target audience are people who have some knowledge in Machine Learning, but interested to get a practical understanding of the Reinforcement Learning domain. A reader should be familiar with Python and the basics of deep learning and machine learning. Understanding of statistics and probability will be a plus, but is not absolutely essential for understanding most of the book's material.

▶What this book covers
⦁ Chapter 1, What is Reinforcement Learning?, contains introduction to RL ideas and main formal models.

⦁ Chapter 2, OpenAI Gym, introduces the reader to the practical aspect of RL, using open-source library gym.

⦁ Chapter 3, Deep Learning with PyTorch, gives a quick overview of the PyTorch library.

⦁ Chapter 4, The Cross-Entropy Method, introduces you to one of the simplest methods of RL to give you the feeling of RL methods and problems.

⦁ Chapter 5, Tabular Learning and the Bellman Equation, gives an introduction to the Value-based family of RL methods.

⦁ Chapter 6, Deep Q-Networks, describes DQN, the extension of basic Value-based methods, allowing to solve complicated environment.

⦁ Chapter 7, DQN Extensions, gives a detailed overview of modern extension to the DQN method, to improve its stability and convergence in complex environments.

⦁ Chapter 8, Stocks Trading Using RL, is the first practical project, applying the DQN method to stock trading.

⦁ Chapter 9, Policy Gradients –- An Alternative, introduces another family of RL methods, based on policy learning.

⦁ Chapter 10, The Actor-Critic Method, describes one of the most widely used method in RL.

⦁ Chapter 11, Asynchronous Advantage Actor-Critic, extends Actor-Critic with parallel environment communication, to improve stability and convergence.

⦁ Chapter 12, Chatbots Training with RL, is the second project, showing how to apply RL methods to NLP problems.

⦁ Chapter 13, Web Navigation, is another long project, applying RL to web page navigation, using MiniWoB set of tasks.

⦁ Chapter 14, Continuous Action Space, describes the specifics of environments, using continuous action spaces and various methods.

⦁ Chapter 15, Trust Regions –- TRPO, PPO, and ACKTR, is yet another chapter about continuous action spaces describing "Trust region" set of methods.

⦁ Chapter 16, Black-Box Optimization in RL, shows another set of methods that don't use gradients in explicit form.

⦁ Chapter 17, Beyond Model-Free –- Imagination, introduces model-based approach to RL, using recent research results about imagination in RL.

⦁ Chapter 18, AlphaGo Zero, describes the AlphaGo Zero method applied to game Connect Four.


출판사 서평

▶Editorial Review
The topic of this book is Reinforcement Learning —.which is a subfield of Machine Learning— focusing on the general and challenging problem of learning optimal behavior in complex environment. The learning process is driven only by reward value and observations obtained from the environment. This model is very general and can be applied to many practical situations from playing games to optimizing complex manufacture processes.

Due to flexibility and generality, the field of Reinforcement Learning is developing very quickly and attracts lots of attention both from researchers trying to improve existing or create new methods, as well as from practitioners interested in solving their problems in the most efficient way.

This book was written as an attempt to fill the obvious lack of practical and structured information about Reinforcement Learning methods and approaches. On one hand, there are lots of research activity all around the world, new research papers are being published almost every day, and a large portion of Deep Learning conferences such as NIPS or ICLR is dedicated to RL methods. There are several large research groups focusing on RL methods application in Robotics, Medicine,
multi-agent systems, and others. The information about the recent research is widely available, but is too specialized and abstract to be understandable without serious efforts. Even worse is the situation with the practical aspect of RL application, as it is not always obvious how to make a step from the abstract method described in the mathematical-heavy form in a research paper to a working implementation solving actual problem. This makes it hard for somebody interested in the field to get an intuitive understanding of methods and ideas behind papers and conference talks. There are some very good blog posts about various RL aspects illustrated with working examples, but the limited format of a blog post allows the author to describe only one or two methods without building a complete structured picture and showing how different methods are related to each other. This book is my attempt to address this issue.

Another aspect of the book is its orientation to practice. Every method is implemented for various environments, from very trivial to quite complex. I've tried to make examples clean and easy to understand, which was made possible by the expressiveness and power of PyTorch. On the other hand, complexity and requirements of examples are oriented to RL hobbyists without access to very large computational resources, such as clusters of GPUs or very powerful workstations. This, I believe, will make the fun-filled and exciting RL domain accessible for a much wider audience than just research groups or large AI companies. However, it is still Deep Reinforcement Learning, so, having access to a GPU is highly recommended. Approximately, half of the examples in the book will benefit from running them on GPU. In addition to traditional medium-sized examples of environments used in RL, such as Atari games or continuous control problems, the book contains three chapters (8, 12, and 13) that contain larger projects, illustrating how RL methods could be applied to more complicated environments and tasks. These examples are still not full-sized real-life projects (otherwise they'll occupy a separate book on their own), but just larger problems illustrating how the RL paradigm can be applied to domains beyond well-established benchmarks.

Another thing to note about examples in the first three parts of the book is that I've tried to make examples self-contained and the source code was shown in full. Sometimes this led to repetition of code pieces (for example, training loop is very similar in most of the methods), but I believe that giving you the freedom to jump directly into the method you want to learn is more important than avoiding few repetitions. All examples in the book is available on Github: https://github.com/ PacktPublishing/Deep-Reinforcement-Learning-Hands-On, and you're welcome to fork them, experiment, and contribute.


저자 소개

⦁ Maxim Lapan
Maxim Lapan is a deep learning enthusiast and independent researcher. His background and 15 years' work expertise as a software developer and a systems architect lays from low-level Linux kernel driver development to performance optimization and design of distributed applications working on thousands of servers. With vast work experiences in big data, Machine Learning, and large parallel distributed HPC and nonHPC systems, he has a talent to explain a gist of complicated things in simple words and vivid examples. His current areas of interest lie in practical applications of Deep Learning, such as Deep Natural Language Processing and Deep Reinforcement Learning. Maxim lives in Moscow, Russian Federation, with his family, and he works for an Israeli start-up as a Senior NLP developer.

목차

▶TABLE of CONTENTS
1: WHAT IS REINFORCEMENT LEARNING?
2: OPENAI GYM
3: DEEP LEARNING WITH PYTORCH
4: THE CROSS-ENTROPY METHOD
5: TABULAR LEARNING AND THE BELLMAN EQUATION
6: DEEP Q-NETWORKS
7: DQN EXTENSIONS
8: STOCKS TRADING USING RL
9: POLICY GRADIENTS – AN ALTERNATIVE
10: THE ACTOR-CRITIC METHOD
11: ASYNCHRONOUS ADVANTAGE ACTOR-CRITIC
12: CHATBOTS TRAINING WITH RL
13: WEB NAVIGATION
14: CONTINUOUS ACTION SPACE
15: TRUST REGIONS – TRPO, PPO, AND ACKTR
16: BLACK-BOX OPTIMIZATION IN RL
17: BEYOND MODEL-FREE – IMAGINATION
18: ALPHAGO ZERO


리뷰

구매자 별점

0.0

점수비율
  • 5
  • 4
  • 3
  • 2
  • 1

0명이 평가함

리뷰 작성 영역

이 책을 평가해주세요!

내가 남긴 별점 0.0

별로예요

그저 그래요

보통이에요

좋아요

최고예요

별점 취소

구매자 표시 기준은 무엇인가요?

'구매자' 표시는 리디에서 유료도서 결제 후 다운로드 하시거나 리디셀렉트 도서를 다운로드하신 경우에만 표시됩니다.

무료 도서 (프로모션 등으로 무료로 전환된 도서 포함)
'구매자'로 표시되지 않습니다.
시리즈 도서 내 무료 도서
'구매자’로 표시되지 않습니다. 하지만 같은 시리즈의 유료 도서를 결제한 뒤 리뷰를 수정하거나 재등록하면 '구매자'로 표시됩니다.
영구 삭제
도서를 영구 삭제해도 ‘구매자’ 표시는 남아있습니다.
결제 취소
‘구매자’ 표시가 자동으로 사라집니다.

이 책과 함께 구매한 책


이 책과 함께 둘러본 책



본문 끝 최상단으로 돌아가기

spinner
모바일 버전