Concept Graph & Resume using Claude 3 Opus | Chat GPT4o | Llama 3:
Resume:
1.- Few-shot classification: Computing classification models that generalize to unseen test sets, given few training samples per category.
2.- Prototypical networks: Embeds training samples, computes class prototypes by averaging, classifies test examples based on nearest prototype.
3.- Meta-learning objective: Learning feature embeddings that generalize well across tasks when used with nearest class prototype classifier.
4.- Meta-learning process: Compute classification model, calculate meta loss measuring generalization error, meta-learn embedding by backpropagating error signal across tasks.
5.- Linear predictors (SVM, logistic/ridge regression): Proposed for computing classification model, incorporated convex optimizer in deep network to solve.
6.- Advantages of SVM over nearest neighbor: Adaptive (task-dependent inference-time adaptation), scalable (less overfitting with larger embeddings, outperforms nearest neighbor in high dimensions).
7.- Gradient computation for SVM: Obtained closed-form gradient expression for embedding network without differentiating through optimization, by leveraging convex nature of problem.
8.- Dual formulation: Addressed computational issues with large embedding dimensions by solving dual problem, expressing model as linear combination of training embeddings.
9.- Results on miniImageNet, tieredImageNet: MetaOpNet improves prototypical network accuracy by ~3% with 30-50% inference time increase. Ridge regression variant is comparable.
10.- Results on CIFAR-FS, FC100: Similar performance on CIFAR-FS, 3% improvement over prototypical networks on harder FC100 dataset with larger train/test class gaps.
11.- Influence of meta-training shot: Model performance generally increases with more meta-training shots, enabling one-time high-shot training for all meta-test scenarios.
Knowledge Vault built byDavid Vivancos 2024