Application of Fuzzy State Aggregation and Policy Hill Climbing to Multi-Agent Systems in Stochastic Environments / Najlacnejšie knihy
Application of Fuzzy State Aggregation and Policy Hill Climbing to Multi-Agent Systems in Stochastic Environments

Kod: 08250540

Application of Fuzzy State Aggregation and Policy Hill Climbing to Multi-Agent Systems in Stochastic Environments

Autor Dean C Wardell

Reinforcement learning is one of the more attractive machine learning technologies, due to its unsupervised learning structure and ability to continually learn even as the operating environment changes. Applying this learning to m ... więcej

56.97

Zwykle: 58.11 €

Oszczędzasz 1.14 €


Dostępna u dostawcy
Wysyłamy za 15 - 20 dni
Dodaj do schowka

Zobacz książki o podobnej tematyce

Podaruj tę książkę jeszcze dziś
  1. Zamów książkę i wybierz "Wyślij jako prezent".
  2. Natychmiast wyślemy Ci bon podarunkowy, który możesz przekazać adresatowi prezentu.
  3. Książka zostanie wysłana do adresata, a Ty o nic nie musisz się martwić.

Dowiedz się więcej

Więcej informacji o Application of Fuzzy State Aggregation and Policy Hill Climbing to Multi-Agent Systems in Stochastic Environments

Za ten zakup dostaniesz 143 punkty

Opis

Reinforcement learning is one of the more attractive machine learning technologies, due to its unsupervised learning structure and ability to continually learn even as the operating environment changes. Applying this learning to multiple cooperative software agents (a multi-agent system) not only allows each individual agent to learn from its own experience, but also opens up the opportunity for the individual agents to learn from the other agents in the system, thus accelerating the rate of learning. This research presents the novel use of fuzzy state aggregation, as the means of function approximation, combined with the fastest policy hill climbing methods of Win or Lose Fast (WoLF) and policy-dynamics based WoLF (PD-WoLF). The combination of fast policy hill climbing and fuzzy state aggregation function approximation is tested in two stochastic environments; Tileworld and the simulated robot soccer domain, RoboCup. The Tileworld results demonstrate that a single agent using the combination of FSA and PHC learns quicker and performs better than combined fuzzy state aggregation and Q-learning reinforcement learning alone. Results from the multi-agent RoboCup domain again illustrate that the policy hill climbing algorithms perform better than Q-learning alone in a multi-agent environment. The learning is further enhanced by allowing the agents to share their experience through a weighted strategy sharing.

Szczegóły książki

Kategoria Książki po angielsku Society & social sciences Education

56.97

Ulubione w innej kategorii



Osobní odběr Bratislava a 2642 dalších

Copyright ©2008-24 najlacnejsie-knihy.sk Wszelkie prawa zastrzeżonePrywatnieCookies


Konto: Logowanie
Všetky knihy sveta na jednom mieste. Navyše za skvelé ceny.

Nákupní košík ( prázdný )

Nakupte za 59,99 € a
máte doručení zdarma.

Twoja lokalizacja: