<?xml version="1.0" encoding="UTF-8"?><xml><records><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Sai Krishna Gottipati</style></author><author><style face="normal" font="default" size="100%">Yashaswi Pathak</style></author><author><style face="normal" font="default" size="100%">Rohan Nuttall</style></author><author><style face="normal" font="default" size="100%">Sahir</style></author><author><style face="normal" font="default" size="100%">Raviteja Chunduru</style></author><author><style face="normal" font="default" size="100%">Ahmed Touati</style></author><author><style face="normal" font="default" size="100%">Sriram Ganapathi Subramanian</style></author><author><style face="normal" font="default" size="100%">Matthew E. Taylor</style></author><author><style face="normal" font="default" size="100%">Sarath Chandar</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Maximum Reward Formulation In Reinforcement Learning</style></title><secondary-title><style face="normal" font="default" size="100%">Deep Reinforcement Learning Workshop</style></secondary-title></titles><dates><year><style  face="normal" font="default" size="100%">2020</style></year></dates><urls><web-urls><url><style face="normal" font="default" size="100%">https://arxiv.org/pdf/2010.03744.pdf</style></url></web-urls></urls><publisher><style face="normal" font="default" size="100%">NeurIPS 2020</style></publisher><language><style face="normal" font="default" size="100%">eng</style></language></record></records></xml>