Communities

Writing
Writing
Codidact Meta
Codidact Meta
The Great Outdoors
The Great Outdoors
Photography & Video
Photography & Video
Scientific Speculation
Scientific Speculation
Cooking
Cooking
Electrical Engineering
Electrical Engineering
Judaism
Judaism
Languages & Linguistics
Languages & Linguistics
Software Development
Software Development
Mathematics
Mathematics
Christianity
Christianity
Code Golf
Code Golf
Music
Music
Physics
Physics
Linux Systems
Linux Systems
Power Users
Power Users
Tabletop RPGs
Tabletop RPGs
Community Proposals
Community Proposals
tag:snake search within a tag
answers:0 unanswered questions
user:xxxx search by author id
score:0.5 posts with 0.5+ score
"snake oil" exact phrase
votes:4 posts with 4+ votes
created:<1w created < 1 week ago
post_type:xxxx type of post
Search help
Notifications
Mark all as read See all your notifications »
Q&A

What would a near-human level AI still struggle at?

+0
−0

Plenty of stories have done AI's before, but many of them feel unrealistic. If you make an AI that is as smart as humans, but with faster reaction speed and ability to process and calculate certain concepts at speeds immensely greater then humans it makes humans rather pointless. The question I'm always left with is why the humans are involved in the stories at all, why not tell the AI what to do and let them do things? Doesn't a human interface just slow things down?

So, I want to imagine a world that has intelligent AI, but are not inherently superior to humans. Imagine a situation where you have a small team of robots and humans involved in some complex activity, something physically challenging, requiring long term planning, adaptation, and quick reflexes. How could I make a system where the humans and AI are both equally relevant in accomplishing the task?

I assume the trick would be to prevent AI from having quite human-level intelligence, but what would an AI like that look like? Assuming the AI is capable of basic conversations, parsing of sentences and communicating at a level of a young child at minimum, and has available to it large processing speeds and memory, what would it struggle at? What are human brains better suited to then any silicone AI we may design in the near future? Would these AI lack ability to plan into the distant future because they still work on the "process every possible outcome one iteration at a time" approach? Would they lack more in creativity etc?

What would such an AI do substantially better then us?

I would be interested in hearing not just what limits they may have, but also your thoughts for the causes of the limits, why is the implementation of the AI we have less capable of doing a given task then a standard human?

History
Why does this post require attention from curators or moderators?
You might want to add some details to your flag.
Why should this post be closed?

This post was sourced from https://worldbuilding.stackexchange.com/q/7105. It is licensed under CC BY-SA 3.0.

0 comment threads

0 answers

Sign up to answer this question »