<?xml version="1.0" encoding="utf-8"?>
<?xml-stylesheet type="text/xsl" href="style/detail_T.xsl"?>
<bibitem type="J">   <ARLID>0502902</ARLID> <utime>20240903170642.1</utime><mtime>20190314235959.9</mtime>   <SCOPUS>85064190063</SCOPUS> <WOS>000457070200009</WOS>  <DOI>10.14736/kyb-2018-6-1218</DOI>           <title language="eng" primary="1">Risk-sensitive Average Optimality in Markov Decision Processes</title>  <specification> <page_count>13 s.</page_count> <media_type>P</media_type> </specification>   <serial><ARLID>cav_un_epca*0297163</ARLID><ISSN>0023-5954</ISSN><title>Kybernetika</title><part_num/><part_title/><volume_id>54</volume_id><volume>6 (2018)</volume><page_num>1218-1230</page_num><publisher><place/><name>Ústav teorie informace a automatizace AV ČR, v. v. i.</name><year/></publisher></serial>    <keyword>controlled Markov processes</keyword>   <keyword>risk-sensitive average optimality</keyword>   <keyword>asymptotic behavior</keyword>    <author primary="1"> <ARLID>cav_un_auth*0101196</ARLID> <full_dept>Department of Econometrics</full_dept>  <share>100%</share> <name1>Sladký</name1> <name2>Karel</name2> <institution>UTIA-B</institution> <full_dept language="cz">Ekonometrie</full_dept> <full_dept language="eng">Department of Econometrics</full_dept> <department language="cz">E</department> <department language="eng">E</department> <garant>K</garant> <fullinstit>Ústav teorie informace a automatizace AV ČR, v. v. i.</fullinstit> </author>   <source> <url>http://library.utia.cas.cz/separaty/2019/E/sladky-0502902.pdf</url> </source>        <cas_special> <project> <ARLID>cav_un_auth*0363963</ARLID> <project_id>GA18-02739S</project_id> <agency>GA ČR</agency> </project>  <abstract language="eng" primary="1">In this note attention is focused on  finding policies optimizing risk-sensitive optimality criteria in Markov decision chains. To this end we assume that the total reward generated by the Markov process is evaluated by an exponential utility function with a given risk-sensitive coefficient. The ratio of the first two moments depends on the value of the risk-sensitive coefficient, if the risk-sensitive coefficient is equal to zero we speak on risk-neutral models. Observe that the  first moment of the generated reward corresponds to the expectation of the total reward and the second central moment of the reward variance. For communicating Markov processes and for some specific classes of unichain processes long run risk-sensitive average reward is independent of the starting state. In this note we present necessary and sufficient condition for existence of optimal policies independent of the starting state in unichain models and characterize the class of average risk-sensitive optimal policies.</abstract>    <action target="CST"> <ARLID>cav_un_auth*0373581</ARLID> <name>Mathematical Methods in Economy and Industry 2017</name>  <dates>20170904</dates> <unknown tag="mrcbC20-s">20170906</unknown> <place>Jindřichův Hradec</place> <country>CZ</country>  </action>  <result_subspec>WOS</result_subspec> <RIV>BB</RIV> <FORD0>10000</FORD0> <FORD1>10100</FORD1> <FORD2>10103</FORD2>    <reportyear>2019</reportyear>      <num_of_auth>1</num_of_auth>  <inst_support> RVO:67985556 </inst_support>  <permalink>http://hdl.handle.net/11104/0295273</permalink>   <confidential>S</confidential>  <unknown tag="mrcbC86"> 1 Article|Proceedings Paper Computer Science Cybernetics </unknown>         <unknown tag="mrcbT16-e">COMPUTERSCIENCE.CYBERNETICS</unknown> <unknown tag="mrcbT16-f">0.591</unknown> <unknown tag="mrcbT16-g">0.155</unknown> <unknown tag="mrcbT16-h">13</unknown> <unknown tag="mrcbT16-i">0.00068</unknown> <unknown tag="mrcbT16-j">0.174</unknown> <unknown tag="mrcbT16-k">891</unknown> <unknown tag="mrcbT16-s">0.268</unknown> <unknown tag="mrcbT16-5">0.500</unknown> <unknown tag="mrcbT16-6">71</unknown> <unknown tag="mrcbT16-7">Q4</unknown> <unknown tag="mrcbT16-B">15.991</unknown> <unknown tag="mrcbT16-C">6.5</unknown> <unknown tag="mrcbT16-D">Q4</unknown> <unknown tag="mrcbT16-E">Q3</unknown> <unknown tag="mrcbT16-M">0.17</unknown> <unknown tag="mrcbT16-N">Q4</unknown> <unknown tag="mrcbT16-P">6.522</unknown> <arlyear>2018</arlyear>       <unknown tag="mrcbU14"> 85064190063 SCOPUS </unknown> <unknown tag="mrcbU24"> PUBMED </unknown> <unknown tag="mrcbU34"> 000457070200009 WOS </unknown> <unknown tag="mrcbU63"> cav_un_epca*0297163 Kybernetika 0023-5954 Roč. 54 č. 6 2018 1218 1230 Ústav teorie informace a automatizace AV ČR, v. v. i. </unknown> </cas_special> </bibitem>