Вы находитесь на странице: 1из 2

Nonetheless, there ought to be the most optimal translation for

a given situation as expressed by metrics such as readability of


the target group, persuasiveness or information density. By
optimizing to those metrics though e.g. Reinforcement Learning
instead of simply imitating human behavior as being practiced
in supervised learning, we’re following a more end-to-end
approach increasing the capabilities of our system considerably.

In this story, we’ll first cover a little bit of a context and first
attempts of leveraging an AI-automation of complex tasks like
self-driving cars, intelligent sorting robots, etc. through
Imitation Learning. Then we’ll stick deeper into why
Reinforcement Learning might be a more capable choice, when
it’s worth to be considered and how it actually works in
principle but also by sketching two popular algorithms
abstractly. Summing up with encouragement on my behalf
about seeking practical use cases of this powerful technique,
does help to spread RL more over right into the industry where
those algorithms support finding more beneficial solutions. Or a
nerd would prefer saying:
from __future__ import
ultimate_ai_solution

Imitation Learning
DAVE-2 by NVIDIA on End to End Learning for Self-Driving Cars

Back in 2016, NVIDIA achieved ground truthing innovation:


Instead of an explicit decomposition of the self-driving car
objective (e.g. lane marking detection, path planning, etc.) in a
hand-crafted internal representation, their system abstracts and
optimizes all necessary processing steps including the internal
representation automatically. In NVIDIA’s approach, the car’s
reaction given certain feedback of its sensors (e.g. steering
angle, speed, front image, etc.), which could, of course,
preprocessed by advanced deep learning techniques, wasn’t
explicitly programmed, alternatively they followed the concept
of Imitation Learning.

Вам также может понравиться