{"id":4303,"date":"2023-08-15T20:09:02","date_gmt":"2023-08-15T20:09:02","guid":{"rendered":"https:\/\/quantil.co\/?post_type=blog&#038;p=4303"},"modified":"2023-08-15T20:18:26","modified_gmt":"2023-08-15T20:18:26","slug":"evaluacion-de-politicas-bajo-ruido-markoviano-mediante-el-algoritmo-de-online-bootstrap-inference","status":"publish","type":"blog","link":"https:\/\/quantil.co\/en\/blog\/evaluacion-de-politicas-bajo-ruido-markoviano-mediante-el-algoritmo-de-online-bootstrap-inference\/","title":{"rendered":"Evaluaci\u00f3n De Pol\u00edticas Bajo Ruido Markoviano Mediante  El Algoritmo De Online Bootstrap Inference"},"excerpt":{"rendered":"<p>Imag\u00ednese poder abstraer el mundo de tal forma que sea posible evaluar cuantitativamente el beneficio que se obtiene de tomar ciertas acciones a lo largo del tiempo. La buena noticia es que esto no es algo descabellado, de hecho una de las maneras de hacerlo es usando la teor\u00eda alrededor del Aprendizaje Reforzado (RL). <\/p>\n","protected":false},"featured_media":4306,"template":"","format":"standard","blog-categories":[28],"blog-tags":[19],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/quantil.co\/en\/wp-json\/wp\/v2\/blog\/4303"}],"collection":[{"href":"https:\/\/quantil.co\/en\/wp-json\/wp\/v2\/blog"}],"about":[{"href":"https:\/\/quantil.co\/en\/wp-json\/wp\/v2\/types\/blog"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/quantil.co\/en\/wp-json\/wp\/v2\/media\/4306"}],"wp:attachment":[{"href":"https:\/\/quantil.co\/en\/wp-json\/wp\/v2\/media?parent=4303"}],"wp:term":[{"taxonomy":"blog-categories","embeddable":true,"href":"https:\/\/quantil.co\/en\/wp-json\/wp\/v2\/blog-categories?post=4303"},{"taxonomy":"blog-tags","embeddable":true,"href":"https:\/\/quantil.co\/en\/wp-json\/wp\/v2\/blog-tags?post=4303"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}