Communities

Writing
Writing
Codidact Meta
Codidact Meta
The Great Outdoors
The Great Outdoors
Photography & Video
Photography & Video
Scientific Speculation
Scientific Speculation
Cooking
Cooking
Electrical Engineering
Electrical Engineering
Judaism
Judaism
Languages & Linguistics
Languages & Linguistics
Software Development
Software Development
Mathematics
Mathematics
Christianity
Christianity
Code Golf
Code Golf
Music
Music
Physics
Physics
Linux Systems
Linux Systems
Power Users
Power Users
Tabletop RPGs
Tabletop RPGs
Community Proposals
Community Proposals
tag:snake search within a tag
answers:0 unanswered questions
user:xxxx search by author id
score:0.5 posts with 0.5+ score
"snake oil" exact phrase
votes:4 posts with 4+ votes
created:<1w created < 1 week ago
post_type:xxxx type of post
Search help
Notifications
Mark all as read See all your notifications »
Q&A

What goals to give to a desirable optimizer AI?

+0
−0

Most of people knowing a bit about AI know the concept of paper-clipper, an AI that was meant to produce more paperclips and eventually turn the entire universe into paperclips at maximum efficiency, first obliterating the humanity as an obstacle in that goal (resisting having their world turned into paperclips).

Let's then not be utter idiots with a goal to have infinite paperclips or something equally useless (convert entire planet into surveilance equipment to monitor the enemy territory better...) and let's design an AI that the humanity would be happy with. A nebulous concept, but I want to keep it so nebulous, because honestly, the AI would probably invent better specific goals than we ever could.

What rules would these be?

  • Serve the common goals and desires of the humanity?
  • Protect life, but don't restrict the freedom in the name of protection?

  • Maybe a simple and abstract one: Act to satisfaction of the humanity ?

Oh, and before we fall into the "literal genie" pitfall, no, globally modifying human brain, so that it perceives happiness about the AI at all times, is not a satisfactory outcome. I believe this precondition could be phrased as a hypothetical conditional: "If the creators of the AI - or their descendants - fully knew these results beforehand, they'd approve."

(so - the AI's meddling with human body/brain/mind structure would be only acceptable within limits that we today would find acceptable - even if the future outlook of the humanity gets more liberal.)

How would one phrase these rules? Something that wouldn't run away into another paper-clipper or shut itself down due to inability to act, or "optimize the humanity away", say, reducing it to a single insane specimen being perfectly happy with a total wipe-out? Or designing the entire future universe into a physical representation of a memetic bomb, image, description of which, if given to any of the human ancestors, would wrap their mind into insanity and make them drool in happy bliss, thus satisfying the need for their (hypothetical) approval.

History
Why does this post require moderator attention?
You might want to add some details to your flag.
Why should this post be closed?

This post was sourced from https://worldbuilding.stackexchange.com/q/17249. It is licensed under CC BY-SA 3.0.

0 comment threads

0 answers

Sign up to answer this question »