31 Matching Annotations
  1. Jun 2025
  2. Mar 2025
  3. Feb 2025
  4. Aug 2024
  5. Jan 2024
  6. Nov 2023
  7. Oct 2023
    1. (Chen, NeurIPS, 2021) Che1, Lu, Rajeswaran, Lee, Grover, Laskin, Abbeel, Srinivas, and Mordatch. "Decision Transformer: Reinforcement Learning via Sequence Modeling". Arxiv preprint rXiv:2106.01345v2, June, 2021.

      Quickly a very influential paper with a new idea of how to learn generative models of action prediction using SARSA training from demonstration trajectories. No optimization of actions or rewards, but target reward is an input.

  8. Jul 2023
  9. Jun 2023
  10. Apr 2023
  11. Jan 2023
  12. Dec 2022
  13. Nov 2022
    1. we propose the Transformer, a model architecture eschewing recurrence and insteadrelying entirely on an attention mechanism to draw global dependencies between input and output.The Transformer allows for significantly more parallelization a

      Using the attention mechanism to determine global dependencies between input and output instead of using recurrent links to past states. This is the essence of their new idea.

  14. Sep 2022
  15. Feb 2022