<?xml version="1.0" encoding="utf-8"?>
<?xml-stylesheet type="text/xsl" href="style/detail_T.xsl"?>
<bibitem type="J">   <ARLID>0602819</ARLID> <utime>20250317083328.4</utime><mtime>20241216235959.9</mtime>   <SCOPUS>85210284955</SCOPUS> <WOS>001370660700039</WOS>  <DOI>10.1109/ACCESS.2024.3497589</DOI>           <title language="eng" primary="1">Knowledge Transfer in Deep Reinforcement Learning via an RL-Specific GAN-Based Correspondence Function</title>  <specification> <page_count>15 s.</page_count> <media_type>E</media_type> </specification>   <serial><ARLID>cav_un_epca*0461036</ARLID><ISSN>2169-3536</ISSN><title>IEEE Access</title><part_num/><part_title/><volume_id>12</volume_id><volume>1 (2024)</volume><page_num>177204-177218</page_num><publisher><place/><name>Institute of Electrical and Electronics Engineers</name><year/></publisher></serial>    <keyword>Deep learning</keyword>   <keyword>Markov decision process</keyword>   <keyword>reinforcement learning</keyword>   <keyword>transfer learning</keyword>   <keyword>knowledge transfer</keyword>    <author primary="1"> <ARLID>cav_un_auth*0333672</ARLID> <name1>Ruman</name1> <name2>Marko</name2> <institution>UTIA-B</institution> <full_dept language="cz">Adaptivní systémy</full_dept> <full_dept language="eng">Department of Adaptive Systems</full_dept> <department language="cz">AS</department> <department language="eng">AS</department> <country>SK</country>  <garant>K</garant> <fullinstit>Ústav teorie informace a automatizace AV ČR, v. v. i.</fullinstit> </author> <author primary="0"> <ARLID>cav_un_auth*0101092</ARLID> <name1>Guy</name1> <name2>Tatiana Valentine</name2> <institution>UTIA-B</institution> <full_dept language="cz">Adaptivní systémy</full_dept> <full_dept>Department of Adaptive Systems</full_dept> <department language="cz">AS</department> <department>AS</department> <full_dept>Department of Adaptive Systems</full_dept> <fullinstit>Ústav teorie informace a automatizace AV ČR, v. v. i.</fullinstit> </author>   <source> <url>https://library.utia.cas.cz/separaty/2024/AS/guy-0602819.pdf</url> </source> <source> <url>https://ieeexplore.ieee.org/document/10752398</url>  </source>        <cas_special> <project> <project_id>CA21169</project_id> <agency>EU-COST</agency> <country>XE</country> <ARLID>cav_un_auth*0452289</ARLID> </project>  <abstract language="eng" primary="1">Deep reinforcement learning has demonstrated superhuman performance in complex decision-making tasks, but it struggles with generalization and knowledge reuse—key aspects of true intelligence. This article introduces a novel approach that modifies Cycle Generative Adversarial Networks specifically for reinforcement learning, enabling effective one-to-one knowledge transfer between two tasks. Our method enhances the loss function with two new components: model loss, which captures dynamic relationships between source and target tasks, and Q-loss, which identifies states significantly influencing the target decision policy. Tested on the 2-D Atari game Pong, our method achieved 100% knowledge transfer in identical tasks and either 100% knowledge transfer or a 30% reduction in training time for a rotated task, depending on the network architecture. In contrast, using standard Generative Adversarial Networks or Cycle Generative Adversarial Networks led to worse performance than training from scratch in the majority of cases. The results demonstrate that the proposed method ensured enhanced knowledge generalization in deep reinforcement learning.</abstract>     <result_subspec>WOS</result_subspec> <RIV>IN</RIV> <FORD0>10000</FORD0> <FORD1>10200</FORD1> <FORD2>10201</FORD2>    <reportyear>2025</reportyear>      <num_of_auth>2</num_of_auth>  <unknown tag="mrcbC52"> 2 4 R hod 4 4rh 4 20250310143107.1 20250310150006.2 </unknown> <inst_support> RVO:67985556 </inst_support>  <permalink>https://hdl.handle.net/11104/0360153</permalink>  <cooperation> <ARLID>cav_un_auth*0478849</ARLID> <name>Provozně ekonomická fakulta, Česká zemědělská univerzita v Praze</name> <institution>PEF CZU</institution> <country>CZ</country> </cooperation>  <confidential>S</confidential>  <unknown tag="mrcbC91"> A </unknown>         <unknown tag="mrcbT16-e">TELECOMMUNICATIONS|ENGINEERING.ELECTRICAL&amp;ELECTRONIC|COMPUTERSCIENCE.INFORMATIONSYSTEMS</unknown> <unknown tag="mrcbT16-f">3.9</unknown> <unknown tag="mrcbT16-g">0.8</unknown> <unknown tag="mrcbT16-h">4.1</unknown> <unknown tag="mrcbT16-i">0.3457</unknown> <unknown tag="mrcbT16-j">0.67</unknown> <unknown tag="mrcbT16-k">294150</unknown> <unknown tag="mrcbT16-q">290</unknown> <unknown tag="mrcbT16-s">0.849</unknown> <unknown tag="mrcbT16-y">50.78</unknown> <unknown tag="mrcbT16-x">5.31</unknown> <unknown tag="mrcbT16-3">185847</unknown> <unknown tag="mrcbT16-4">Q1</unknown> <unknown tag="mrcbT16-5">3.200</unknown> <unknown tag="mrcbT16-6">13193</unknown> <unknown tag="mrcbT16-7">Q2</unknown> <unknown tag="mrcbT16-C">62.3</unknown> <unknown tag="mrcbT16-M">0.83</unknown> <unknown tag="mrcbT16-N">Q2</unknown> <unknown tag="mrcbT16-P">64.8</unknown> <arlyear>2024</arlyear>    <unknown tag="mrcbTft">  Soubory v repozitáři: guy-0602819.pdf </unknown>    <unknown tag="mrcbU14"> 85210284955 SCOPUS </unknown> <unknown tag="mrcbU24"> PUBMED </unknown> <unknown tag="mrcbU34"> 001370660700039 WOS </unknown> <unknown tag="mrcbU63"> cav_un_epca*0461036 IEEE Access Roč. 12 č. 1 2024 177204 177218 2169-3536 2169-3536 Institute of Electrical and Electronics Engineers </unknown> </cas_special> </bibitem>