Concept Graph & Resume using Claude 3 Opus | Chat GPT4 | Gemini Adv | Llama 3:
Resume:
1.-Dataset bias is an issue where datasets do not adequately cover rare situations or have uneven representation of various attributes.
2.-Collecting more data to address bias can be very difficult and costly in practice due to exponential growth in labeling budget.
3.-Domain adaptation is the problem of adapting a model trained on a source domain to perform well on a target domain.
4.-The source domain has labeled data while the target domain is unlabeled, and there is a distributional difference between them.
5.-Poor model performance under distribution shift is caused by differences in the distribution of training and test data points.
6.-Domain confusion aligns the source and target distributions by adding an unsupervised loss to encourage similar statistics across domains.
7.-Adversarial domain alignment uses a domain discriminator network to distinguish domains while the encoder tries to confuse it, aligning distributions.
8.-Pixel-level domain adaptation uses conditional GANs to translate source images to match the style of the target domain.
9.-Few-shot pixel-space translation, like COCO-FUNIT, translates a source image to a target style given a few target examples.
10.-Pixel alignment makes adaptation effects more interpretable but can have GAN issues; feature alignment is more flexible but can fail silently.
11.-Class-conditional alignment uses the task classifier loss to align target features with source and push them from decision boundaries.
12.-Open set domain adaptation handles differing label spaces between source and target by detecting and rejecting unknown target classes.
13.-DANCE clusters target data, aligns known classes with source, and pushes unknown target classes away via entropy separation loss.
14.-Cross-domain self-supervision finds nearest neighbors between domains to align representations without labels as an unsupervised pre-training step.
15.-Datasets with multiple distributions like Office-31, VisDA, and DomainNet enable benchmarking progress on adaptation algorithms.
16.-It's difficult to have an unbiased dataset, as equal coverage of all variations and latent factors in complex data is hard.
17.-Dataset collection takes shortcuts for cost/time, like web scraping, which introduces biases compared to more controlled gathering.
18.-Some adaptation techniques like distribution alignment losses are applicable beyond vision to domains like text that face similar shifts.
19.-In NLP, news data faces domain shifts due to differences in political leaning and format of news sources.
20.-Handling unknown domains at test time without target data is much harder and requires learning to generalize from diverse training domains.
21.-For time series data, domains may change continuously over time, requiring methods that can adapt in an online fashion.
22.-Applying adaptation methods in practice faces challenges in hyperparameter tuning when target labels are unavailable to guide choices.
23.-Some overlap exists between dataset bias/shift issues and fairness issues, as both relate to performance gaps across data attributes.
24.-However, fairness involves additional considerations beyond accuracy, like equalizing error rates, that require more nuance than pure adaptation.
25.-Making decisions about people requires carefully evaluating and understanding performance on target distributions, not just naive adaptation.
26.-Recent progress in unsupervised learning could help address domain shift by learning from unlabeled data across multiple domains.
27.-Open problems remain in handling unknown class overlap between domains and generalizing to fully unseen domains without target data.
28.-A visual domain adaptation competition called VisDA will be featured at NeurIPS 2021, focused on the universal adaptation setting.
29.-Research interests include the intersection of open set recognition, domain shift, fairness, and learning from unlabeled multi-domain data.
30.-The speaker invites further discussion on these topics and will check the chat for any additional questions.
Knowledge Vault built byDavid Vivancos 2024