Concept Graph & Resume using Claude 3.5 Sonnet | Chat GPT4o | Llama 3:
Resume:
1.- Jason Weston got PhD in 2000, co-supervised by Vapnik. Known for work on SVMs, NLP with neural nets, memory nets.
2.- Goal is to build intelligent conversational agent that can learn from dialogue. Challenges include reasoning, long conversations, learning new knowledge.
3.- Memory networks combine large memory with learning component that can read/write to memory. Many possible variations in architecture.
4.- Toy tasks designed to test reasoning capabilities needed for dialogue, like tracking location of objects, counting, deduction, pathfinding.
5.- First memory network used hard attention over memories to find supporting facts, trained with supporting facts as additional supervision.
6.- Increasing memory hops improves performance on toy tasks requiring multiple supporting facts. Some tasks still challenging, like pathfinding.
7.- End-to-end memory networks use continuous attention over memories to train without supervision of supporting facts. Attention is interpretable.
8.- On toy tasks, multiple hops improve accuracy for end-to-end networks, but still fall short of strongly supervised version on some tasks.
9.- Related work includes NTM, stack-augmented RNNs, attention-based models for MT, NLP tasks. RAM workshop at NIPS explores reasoning, attention, memory.
10.- Large language modeling datasets test ability of models to use long-term context. Analysis shows attention hops flip between nearby and faraway words.
11.- New datasets test reasoning over long contexts through cloze-style QA (CBT, CNN/DailyMail). Humans use context to improve accuracy.
12.- Self-supervision on memories (assuming answer is in them) and multi-hop attention help on CBT. Still a gap to human performance.
13.- Memory networks competitive on QA datasets like WebQuestions, WikiQA, but focus has been more on learning algorithms than feature engineering.
14.- Key-value memory networks separate memories into keys for addressing and values for reading. Allows different representations for each to improve performance.
15.- Movie dialogue dataset tests both QA and recommendation abilities in conversations. Baseline models provided, but challenges remain.
16.- Memory networks achieve strong results on Ubuntu dialogue corpus, but best model so far is an RNN-CNN architecture.
17.- More realistic toy dialogue tasks could help drive innovative model architectures. Understanding successes/failures on real data remains challenging.
18.- Supervised datasets exist, but reinforcement learning through interaction may be needed, similar to how children learn language.
19.- Forward prediction of conversational responses provides an alternative training signal to rewards. Textual feedback can be more informative than binary rewards.
20.- Dialogue-based language learning paper proposes architectures and training procedures for learning from various types of interactive feedback without explicit rewards.
21.- Code and data available online for memory networks and related research. Many open questions remain in reasoning, attention and memory.
22.- Motivation is building models that can engage in meaningful dialogue by combining reasoning, attention, and memory.
23.- Attention enables scaling to large memories by retrieving relevant information as needed. Increasing hops allows deeper reasoning.
24.- Self-supervised memory helps performance by assuming answers are present in the input and learning to pick them out.
25.- Separating memory into key/value allows different representations for retrieval and prediction. Improves performance on WikiQA.
26.- Movie dialogue data tests both factual QA and recommendation. Joint model does both but still room for improvement.
27.- Goal is to have one model that can engage in open-ended dialogue, asking and answering questions, making recommendations, etc.
28.- Reinforcement learning from conversational interaction, rather than supervised datasets, may be key to achieving general dialogue agents.
29.- Rich textual feedback provides more than just a reward signal. Predicting feedback trains model to understand answers.
30.- Much future work remains to solve reasoning, attention, memory challenges and build intelligent dialogue agents that can learn.
Knowledge Vault built byDavid Vivancos 2024