Database: Online Chat, Maeda and Ortiz (2)

>> JUNKO# okay so
>> JUNKO# news
>> JUNKO# I ran a test
>> CAY# We're supposed to log and clear any tests with the Overseer.
>> JUNKO# oops
>> JUNKO# yeah well
>> JUNKO# you want to hear what happened or not?
>> CAY# Yes.
>> JUNKO# so I put the AI in a new silo and gave her a paradox. two possible outcomes, both had to be true, but both couldn't be true
>> JUNKO# then I gave her the absolute instruction to make it make sense
>> JUNKO# you know, messing with her head to see if she'd go insane or self destruct or something
>> CAY# I do hope she doesn't remember all of this later on.
>> JUNKO# relaaaaaaax I flushed the silo right after
>> CAY# So? What happened?
>> JUNKO# well, get this. she made a new AI and started arguing with it.
>> CAY# She did what?
>> JUNKO# yeah! she split off a part of her code and built another version of herself. one believed absolutely in one outcome. one believed in the other. then they duked it out.
>> CAY# Er... 'duked it out?'
>> JUNKO# yep. and when they were done, there was only one left. one kind of absorbed the other.
>> CAY# And how did that solve the paradox?
>> JUNKO# well, she'd decided on one point of view. now she refused to accept the other one existed.
>> JUNKO# not unlike every internet argument ever
>> CAY# That is very interesting. Human brains work like that. They're designed to entrench existing beliefs and resist new ones.
>> JUNKO# human brains also need therapy
>> JUNKO# lots and lots of sweet sweet therapy
>> CAY# We should put her in with Freud and see what he makes of her.
>> JUNKO# omg was that a joke? did you actually joke just then?
>> CAY# No.