4 Matching Annotations
- Nov 2022
-
arxiv.org arxiv.org
-
Extractive summarization may be regarded as acontextual bandit as follows. Each document is acontext, and each ordered subset of a document’ssentences is a different action
We can represent extractive summarization as a bandit problem by treating the document as the context and possible reorderings of sentences as actions an agent could take
-
andit is a decision-making formal-ization in which an agent repeatedly chooses oneof several actions, and receives a reward based onthis choice.
Definition for contextual bandit: an agent that repeatedly choses one of several actions and receives a reward based on this choice.
Tags
Annotators
URL
-
-
aclanthology.org aclanthology.org
-
BanditSum a hierarchical bi-LSTM
Banditsum uses bi-directional LSTM encoding. It generates sentence-level representations
Tags
Annotators
URL
-
- Oct 2020
-
reisub0.github.io reisub0.github.io
-
Most people seem to follow one of two strategies - and these strategies come under the umbrella of tree-traversal algorithms in computer science.
Deciding whether you want to go deep into one topic, or explore more topics, can be seen as a choice between two types of tree-traversal algorithms: depth-first and breadth-first.
This also reminds me of the Explore-Exploit problem in machine learning, which I believe is related to the Multi-Armed Bandit Problem.
-