Concept Graph & Resume using Claude 3 Opus | Chat GPT4 | Gemini Adv | Llama 3:
Resume:
1.-Papers using TCAP in medicine and science provide the best evidence, allowing model predictions to align with current medical knowledge and guidelines.
2.-Using concepts familiar to doctors makes the language work for both machine learning researchers and experts in other fields.
3.-The TCAP work is widely popular at Google, highlighted by Sundar Pichai, and won a UNESCO NetExplorer Award for potential impact.
4.-To expand knowledge, examples decompose the embedding space using PCA or clustering to reveal machine concepts expressed in human-understandable ways.
5.-A trained model pays attention to tiles on a platform for one class, and humans holding dumbbells for the dumbbell class.
6.-Measuring completeness of discovered concepts is possible, though machines' concepts may be too wild to express using available images.
7.-The Dissect paper trains a generative model using gradients of a trained classification model to draw the machine's learned concepts.
8.-Limitations exist in current methods to expand knowledge, such as validating new concepts on limited synthetic datasets or with domain experts.
9.-An in-depth study of how the self-trained chess model AlphaZero sees the world provides insight into expanding shared basis with humans.
10.-Human chess concepts like material imbalance and in-check exist in AlphaZero, but when and where they are learned varies.
11.-AlphaZero's chess development differs from humans', with more diverse opening moves and an "aha" moment where skills explode and style emerges.
12.-A tool using non-negative matrix factorization allows exploring AlphaZero's representational space, marking a first step towards many potential follow-up works.
13.-Lack of alignment between humans and machines could inspire human creativity, as explored in an open-source project with designers and artists.
14.-Mood Board Search enables visual dialogue, with humans providing seeding images and machines responding based on their different representational mapping.
15.-Artists found the machine's differing perspective helped them see their own photography in new ways and escape the ordinary.
16.-Concept Camera, another open-sourced app, allows seeing through your camera from someone else's conceptual eyes.
17.-Projects bringing out surprising insights in humans represent a different way to expand knowledge through concept-based human-machine dialogue.
18.-Many collaborators over years contributed to shaping the opportunity to influence human and machine thinking and future relationships through language.
19.-Implications of the work on using visualization to diagnose machine learning errors underscore the need for skepticism and extensive testing.
20.-Parallel efforts between science and engineering, both theoretical and practical, are needed to surface errors and develop interpretability tools.
21.-Collaboration with experts in human psychology is crucial given human biases and the challenge of understanding ourselves as we develop machines.
22.-Balanced consideration of both inherently interpretable models and post-hoc explanation methods is warranted given the current state of knowledge.
23.-Testing explanation methods by intentionally inserting bugs is important to verify they actually detect known problems before practical use.
24.-Saliency maps may reflect data distribution variance rather than prediction-relevant information, so testing is key to determine explanation method fit.
25.-Fundamental differences in how humans and machines perceive pixel-level data may underlie some observed saliency map limitations.
26.-General interpretability tools like TCAP have been successfully applied across diverse data types including language, audio, images, and medical data.
27.-An abstract interpretability language envisioned to enable broad alignment is not yet realized but an aspirational long-term goal.
28.-The language could eventually allow both experts and laypeople to communicate insights to machine learning models in accessible ways.
29.-Potential applications of TCAV to improve NLP model understanding, like for abusive language detection, are an intriguing area of study.
30.-Effective interpretability methods should consider human cognition, like leveraging our quick parsing of visual information or domain-relevant sequential thinking constraints.
Knowledge Vault built byDavid Vivancos 2024