User:IssaRice/AI safety/Comparison of terms related to agency: Difference between revisions

From Machinelearning
No edit summary
(Redirected page to /)
 
(14 intermediate revisions by the same user not shown)
Line 1: Line 1:
{| class="wikitable"
#redirect [[/]]
! Term !! Opposite
|-
| Agent ||
|-
| Optimizer, optimization process ||
|-
| consequentialist ||
|-
| expected utility maximizer ||
|-
| goal-directed, goal-based || act-based?
|-
| pseudoconsequentialist ||
|-
| mesa-optimizer ||
|}


parameters to check for:
This page has moved to https://wiki.issarice.com/wiki/Comparison_of_terms_related_to_agency
 
* is it searching through a list of potential answers?
* does it have an explicit model of the world? i.e. it has counterfactuals (see drescher on subactivation)
* can it be modeled as having a utility function?
* can we take an intentional stance toward it? i.e., is it useful (so far as predicting what it will do is concerned) to model it as having intentions?
* is it solving some sort of optimization problem? (but what counts as an optimization problem?)
* origin: was it itself produced by some sort of optimization process?
** eliezer's GLUT idea of "trace the improbability"
* does it hit a small target, out of a large space of possibilities?
* how many elements in the space of possibilities is it instantiating?
* cost of evaluation of options/how good of a feedback we get for potential outputs
* online vs offline (maybe also one-shot vs continually outputting things)
* how well does it continue working when the environment/input changes?
* coherence (if it makes sense to assign preferences)
 
examples to check against:
 
* humans
* evolution/natural selection
* bottlecap
* RL system playing Pong without an explicit model
* tool AGI/CAIS
* task AGI
* KANSI
* targeting system on a rocket
* single-step filter
* chess-playing algorithm that just does tree search (e.g. alpha-beta pruning algorithm)
* a simple feed-forward neural network (e.g. one that recognizes MNIST digits)
* a thermostat
* a plant
* multi-armed bandit problem
* Solomonoff induction (outer layer/top-level reasoning)
 
==References==
 
https://www.lesswrong.com/posts/ZDZmopKquzHYPRNxq/selection-vs-control
 
https://www.lesswrong.com/posts/rvxcSc6wdcCfaX6GZ/two-senses-of-optimizer

Latest revision as of 01:09, 18 May 2020