-
Notifications
You must be signed in to change notification settings - Fork 6
/
Copy pathCITATION.cff
59 lines (54 loc) · 2.1 KB
/
CITATION.cff
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
# This CITATION.cff file was generated with cffinit.
# Visit https://bit.ly/cffinit to generate yours today!
cff-version: 1.2.0
title: Internet Explorer
message: >-
If you use this software, please cite it using the
metadata from this file.
type: software
authors:
- given-names: Ellis
name-particle: L
family-names: Brown
name-suffix: II
email: [email protected]
affiliation: Carnegie Mellon University
orcid: 'https://orcid.org/0000-0002-8117-0778'
- given-names: Alexander
name-particle: C
family-names: Li
email: [email protected]
affiliation: Carnegie Mellon University
orcid: 'https://orcid.org/0000-0002-9884-6383'
repository-code: 'https://github.com/internet-explorer-ssl/internet-explorer'
url: 'https://internet-explorer-ssl.github.io'
abstract: >-
Vision models heavily rely on fine-tuning general-purpose
models pre-trained on large, static datasets. These
general-purpose models only understand knowledge within
their pre-training datasets, which are tiny, out-of-date
snapshots of the Internet—where billions of images are
uploaded each day.
We suggest an alternate approach: rather than hoping our
static datasets transfer to our desired tasks after
large-scale pre-training, we propose dynamically utilizing
the Internet to quickly train a small-scale model that
does extremely well on the task at hand. Our approach,
called Internet Explorer, explores the web in a
self-supervised manner to progressively find relevant
examples that improve performance on a desired target
dataset. It cycles between searching for images on the
Internet with text queries, self-supervised training on
downloaded images, determining which images were useful,
and prioritizing what to search for next.
We evaluate Internet Explorer across several datasets and
show that it outperforms or matches CLIP oracle
performance by using just a single GPU desktop to actively
query the Internet for 30–40 hours.
keywords:
- Machine Learning
- Computer Vision
- Self-Supervised Learning
- Representation Learning
- Online Learning
license: MIT