使用MATLAB实现增强学习-理解基础知识并建立环境

更新时间:2023-06-03 22:47:48 阅读: 评论:0

guimi
Reinforcement Learning with MATLAB Understanding the Basics and Setting Up
the Environment
What Is Reinforcement Learning?
Reinforcement learning is learning what to do—how to map谷雨是什么季节
situations to actions—so as to maximize a numerical reward
signal. The learner is not told which actions to take, but instead
must discover which actions yield the most reward by trying them.
—Sutton and Barto, Reinforcement Learning: An Introduction
Reinforcement learning (RL) has successfully trained
computer programs to play games at a level higher than the
world’s best human players.
The programs find the best action to take in games with
large state and action spaces, imperfect world information,
and uncertainty around how short-term actions pay off in the
long run.
Engineers face the same types of challenges when designing
controllers for real systems. Can reinforcement learning also
help solve complex control problems like making a robot walk游岳阳楼记
or driving an autonomous car?
This ebook answers that question by explaining what RL is
in the context of traditional control problems and helps you
understand how to t up and solve the RL problem.
The Goal of Control
Broadly speaking, the goal of a control system is to determine the
correct inputs (actions) into a system that will generate the desired
剪鼻毛system behavior.
With feedback control systems, the controller us state obrvations
to improve performance and correct for random disturbances and
errors. Engineers u that feedback, along with a model of the
plant and environment, to design the controller to meet the system
requirements.
qq解除安全模式
This concept is simple to put into words, but it can quickly become
difficult to achieve when the system is hard to model, is highly
nonlinear, or has large state and action spaces.
imagine developing a control system for a walking robot.
To control the robot (i.e., the system), you command potentially dozens
of motors that operate each of the joints in the arms and legs.
Each command is an action you can take. The state obrvations
come from multiple sources, including a camera vision nsor,
accelerometers, gyros, and encoders for each of the motors.小猪奴尼
• Determine the right combination of motor torques to get the robot
walking and keep it balanced.
• Operate in an environment that has random obstacles that
need to be avoided.
• Reject disturbances like wind gusts.
我们的中国心A control system design would need to handle the as well as any
additional requirements like maintaining balance while walking down
a steep hillside or across a patch of ice.
鹌鹑蛋煮几分钟Typically, the best way to approach this problem is to break it up into smaller discrete ctions that can be solved independently.
For example, you could build a process that extracts features from the camera images. The might be things like the location and type of obstacle, or the location of the robot in a global reference fram
e. Combine tho states with the procesd obrvations from the other nsors to complete the full state estimation.
The estimated state and the reference would feed into the controller, which would likely consist of multiple nested control loops. The outer loop would be responsible for managing high-level robot behavior (like maybe maintaining balance), and the inner loops manage low-level behaviors and individual actuators.
All solved? Not quite.
The loops interact with each other, which makes design and tuning challenging. Also, determining the best way to structure the loops and break up the problem is not simple.

本文发布于:2023-06-03 22:47:48,感谢您对本站的认可!

本文链接:https://www.wtabcd.cn/fanwen/fan/82/852867.html

版权声明:本站内容均来自互联网,仅供演示用,请勿用于商业和其他非法用途。如果侵犯了您的权益请与我们联系,我们将在24小时内删除。

标签:解除   小猪   模式   奴尼
相关文章
留言与评论(共有 0 条评论)
   
验证码:
推荐文章
排行榜
Copyright ©2019-2022 Comsenz Inc.Powered by © 专利检索| 网站地图