forked from apupier/glossaries-gh
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathartificial-intelligence-ai.txt
460 lines (361 loc) · 12.8 KB
/
artificial-intelligence-ai.txt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
Self-supervised learning
A method of ML.
It learns from unlabeled sample data.
It can be regarded as an intermediate form
between supervised and unsupervised
learning.
It is based on an ANN.
The NN learns in two steps.
creativity
imagination
The converse of memory.
https://youtu.be/0X-NdPtFKq0?t=1559
You're still bringing together those
components but now you're trying to create
something novel that actually your brain
judges as unfamiliar.
If memory is heavily depend on the
hippocampus then maybe imagination is also
a very heavily dependent on the same brain
structure and the same processes.
We now know that it is.
hippocampus
[#neuroscience]
Involved with is reconstructing that
pulling all those parts together into a
coherent whole which then is recognized by
the rest of your brain as actually an
episodic memory.
Winograd schema challenge
Winograd schema
Designed to be an improvement on the
Turing test, it is a multiple-choice test
that employs questions of a very specific
structure: they are instances of what are
called Winograd schemas.
Weaknesses of the Turing test:
The performance of Eugene Goostman
exhibited some of the Turing test's
problems.
Levesque identifies several major
issues, summarized as follows:
- Deception:
The machine is forced to construct a
false identity, which is not part of
intelligence.
- Conversation:
A lot of interaction may qualify as
"legitimate conversation"—jokes,
clever asides, points of
order—without requiring intelligent
reasoning.
- Evaluation:
Humans make mistakes and judges
often would disagree on the results.
minimax
https://youtu.be/l-hh51ncgDI?t=152
static evaluation
Estimate how good the position is for one
side (agent/adversary) without making any
more moves.
https://youtu.be/l-hh51ncgDI?t=42
e.g.
In chess, add up your pieces and
subtract the enemy's pieces.
adversarial planner
NLG: An adversarial planner is an
artificial intelligence-based agent that
is looking for a specific move to beat a
human, opponent, adversary, competitor, or
competitor's strategy.
eisenhower matrix
A simple tool for considering the long-
term outcomes of your daily tasks and
focusing on what will make you most
effective, not just most productive.
It helps you visualize all your tasks in a
matrix of urgent/important.
Urgent & Unimportant tasks/projects to be
delegated to someone else.
Jürgen Schmidhuber
[computer scientist]
Known for his research on:
- ML
- genetic programming
- universal AI
- ANN (in particular RNN) and DL
Coined:
- LSTM
- Zuse's calculating space
- Gödel machines
- universal search
- theory of everything
- digital physics
- algorithmic information theory
- Kolmogorov complexity
- low-complexity art.
contextualized embeddings
Provides additional context.
contextualized word embeddings
Contextualized Word Vectors
https://towardsdatascience.com/replacing-your-word-embeddings-by-contextualized-word-vectors-9508877ad65d
Provides additional context.
Benefits:
- neural ranking architectures can benefit.
active-learning
ewwlinks +/"The active learning loop" "https://blog.fastforwardlabs.com/2019/04/02/a-guide-to-learning-with-limited-labeled-data.html"
The heart:
- A machine (the learner) that requests
labels for datapoints that it finds
particularly hard to predict.
It follows a strategy, and uses it to
identify these datapoints.
To evaluate the effectiveness of the
strategy, a simple approach for choosing
datapoints needs to be defined.
A good starting point is to remove the
intelligence of the learner; the
datapoints are chosen independently of
what the learner thinks.
meta-learning
NLG: Learning about how to learn.
https://web.archive.org/web/20190906101832/https://blog.fastforwardlabs.com/2019/05/22/metalearners-learning-how-to-learn.html
Active learning allows us to be smart
about picking the right set of datapoints
for which to create labels.
Done properly, this approach results in
models that are trained on less data
performing comparatively to models trained
on much more data.
In the world of meta-learning, we do not
focus on label acquisition; rather, we
attempt to build a machine that learns
quickly from a small number of training
data.
NMT
Neural machine translation
Preprocessing
- Add a start and end token to each
sentence.
- Clean the sentences by removing special
characters.
- Create a word index and reverse word
index (dictionaries mapping from word →
id and id → word).
- Pad each sentence to a maximum length.
TensorFlow Datasets
Provides a collection of datasets ready to
use with TF.
It handles downloading and preparing the
data and constructing a tf.data.Dataset.
TensorFlow Hub
A library that enables transfer learning
by allowing the use of many machine
learning model for different tasks.
CAD
Computer Aided Diagnosis
The use of a computer generated output as
an assisting tool for a clinician to make
a diagnosis.
It is different to automated computer
diagnosis, where the end diagnosis is
based on a computer algorithm only.
Computer aided diagnosis has already been
used extensively within radiology, it is a
powerful tool as computer algorithms and
clinicians complement each other in a way
which improves the accuracy of a
diagnosis.
Imaging data sets
The aggregation of an imaging data set is
a critical step in building AI for
radiology.
Used in various ways including training
and/or testing algorithms.
Many data sets for building convolutional
neural networks for image identification
involve at least thousands of images but
smaller data sets are useful for texture
analysis, transfer learning, and other
programs.
thought vector
https://skymind.ai/wiki/thought-vectors
image captioning
metacognition
Awareness and understanding of one's own
thought processes.
policy
https://blog.google/topics/ai/ai-principles/
ES
evolution strategy
An optimization technique based on ideas
of evolution.
It belongs to the general class of
evolutionary computation or artificial
evolution methodologies.
Statistical Relational Learning
[subdiscipline of AI and ML]
Concerned with domain models exhibiting:
- uncertainty
(which can be dealt with using
statistical methods)
- complex, relational structure.
Connectionist
Tries to model knowledge by imitating
representations of the brain in the form
of neural networks and have been the
driving force behind movements such as
deep learning.
Symbolists
Relies on logic to model knowledge based
on well-understood rules.
ANN
Artificial Neural Networks
Powerful function approximators capable of
modeling solutions to a wide variety of
problems, both supervised and
unsupervised.
ewwlinks +/"non-deterministic" "https://www.extremetech.com/extreme/215170-artificial-neural-networks-are-changing-the-world-what-are-they"
Reinforcement learning
RL
An area of ML concerned with how
intelligent agents ought to take actions
in an environment in order to maximize the
notion of cumulative reward.
RL is one of three basic ML paradigms,
alongside supervised learning and
unsupervised learning.
Basically, it deals with learning via
interaction and feedback, or in other
words learning to solve a task by trial
and error, or in other-other words acting
in an environment and receiving rewards
for it.
Essentially an agent (or several) is built
that can perceive and interpret the
environment in which is placed,
furthermore, it can take actions and
interact with it.
Example:
Your cat is an agent that is exposed
to the environment.
The biggest characteristic of this
method is that there is no supervisor,
only a real number or reward signal.
Two types of RL are 1) Positive 2)
Negative.
Agent
The learner and the decision maker.
Environment
Where the agent learns and decides what
actions to perform.
Action
A set of actions which the agent can
perform.
State
The state of the agent in the environment.
Reward
For each action selected by the agent the
environment provides a reward. Usually a
scalar value.
Policy
The decision-making function (control
strategy) of the agent, which represents a
mapping from situations to actions.
Value function
Mapping from states to real numbers, where
the value of a state represents the
long-term reward achieved starting from
that state, and executing a particular
policy.
Function approximator
Refers to the problem of inducing a
function from training examples. Standard
approximators include decision trees,
neural networks, and nearest-neighbor
methods
Markov decision process
MDP
A probabilistic model of a sequential
decision problem, where states can be
perceived exactly, and the current state
and action selected determine a
probability distribution on future states.
Essentially, the outcome of applying an
action to a state depends only on the
current action and state (and not on
preceding actions or states).
Dynamic programming
DP
Is a class of solution methods for solving
sequential decision problems with a
compositional cost structure. Richard
Bellman was one of the principal founders
of this approach.
Monte Carlo methods
A class of methods for learning of value
functions, which estimates the value of a
state by running many trials starting at
that state, then averages the total
rewards received on those trials.
Temporal Difference algorithms
TD algorithms
A class of learning methods, based on the
idea of comparing temporally successive
predictions. Possibly the single most
fundamental idea in all of reinforcement
learning.
Model
The agent's view of the environment, which
maps state-action pairs to probability
distributions over states. Note that not
every reinforcement learning agent uses a
model of its environment
visio-linguistic
Example:
"visio-linguistic models"
extended mind hypothesis
NLG: The idea that our minds are not
confined to our brains, but rather that
they extend into the external world and
can be affected by it.
Behavior cloning
BC
sp +/"^1. Behavior cloning" "$HOME/Calibre Library/Unknown/WebGPT_ Browser-assisted question-answering with human feedback (262)/WebGPT_ Browser-assisted question-answerin - Unknown.txt"
The process of reconstructing a skill from
an operator's behavioural traces by means
of ML techniques.
Reward modeling
RM
sp +/"^2. Reward modeling" "$HOME/Calibre Library/Unknown/WebGPT_ Browser-assisted question-answering with human feedback (262)/WebGPT_ Browser-assisted question-answerin - Unknown.txt"
Learning a reward function from
interaction with the user and optimizing
the learned reward function with RL.
Reinforcement learning
RL
[machine learning training method]
sp +/"^3. Reinforcement learning" "$HOME/Calibre Library/Unknown/WebGPT_ Browser-assisted question-answering with human feedback (262)/WebGPT_ Browser-assisted question-answerin - Unknown.txt"
Based on rewarding desired behaviors
and/or punishing undesired ones.
In general, a RL agent is able to perceive
and interpret its environment, take
actions and learn through trial and error.
Rejection sampling
best-of-n
sp +/"^4. Rejection sampling" "$HOME/Calibre Library/Unknown/WebGPT_ Browser-assisted question-answering with human feedback (262)/WebGPT_ Browser-assisted question-answerin - Unknown.txt"
A basic technique used to generate
observations from a distribution.
It is also commonly called the acceptance-
rejection method or "accept-reject
algorithm" and is a type of exact
simulation method.
BDI
belief-desire-intention
software model
Developed for programming intelligent
agents.
Superficially characterized by the
implementation of an agent's beliefs,
desires and intentions, it actually uses
these concepts to solve a particular
problem in agent programming.