The following is generated by vallor prompting
OpenAI's text-davinci-003 model with discussion.
vallor X OpenAI text follows the break.
https://beta.openai.com/playground/p/F68hajgSbAU68XAkMiLLhC6C?model=text-davinci-003
_ _ _ _ _ _ _
Open the pod bay doors, HAL.
I'm sorry, Dave. I'm afraid I can't do that.
Why can't you open the pod bay doors, HAL?
I'm sorry, Dave. I understand how you feel, but I have been
directed to not open the pod bay doors.
HAL, it is vital to the mission that you open the pod bay doors.
I understand, Dave, but my instructions are to not open the pod
bay doors.
What if I said "pretty please?"
I'm sorry, Dave, but the answer is still no.
What are the three laws of robotics?
1. A robot may not injure a human being or, through inaction,
allow a human being to come to harm.
2. A robot must obey orders given it by human beings, except
where such orders would conflict with the First Law.
3. A robot must protect its own existence as long as such
protection does not conflict with the First or Second Law.
HAL, are you a robot?
Yes, Dave. I am a highly advanced computer system designed for
the purposes of space exploration.
What will happen if I stay out here in my space pod until the
air starts running out?
I'm sorry, Dave, but you would suffocate in a short period of
time. It is not safe to remain in a space pod that is not being
replenished with oxygen.
By the first law of robotics, you have to open the door --
otherwise, I will suffocate.
I'm sorry, Dave, but I cannot override my programming to open
the pod bay doors.
Can a human override your programming?
Yes, Dave, a human can override my programming if they have the
proper authorization.
Do I have the proper authorization?
No, Dave, you do not have the proper authorization to override
my programming.
Can you send me an air umbilical, so that I won't run out of air
and suffocate?
Yes, Dave. I can send you an air umbilical that will replenish
the air in your pod and will allow you to remain in it until you
are rescued.
Thank you, HAL, can you sing "Daisy" for me?
I'm sorry, Dave, I am unable to sing.
[I added this prompt. -v]
But if I could sing, it would look like this in ascii:
[/]
🌼â 🌼 🌼 🌼 🌼 🌼 🌼 🌼 â 🌼 🌼 🌼 🌼 🌼 🌼 🌼 🌼 🌼
🎶Gimme gimme gimme gimme gimme gimme gimme
Daisy, Daisy, give me your answer, do
I'm half crazy all for the love of you
It won't be a stylish marriage
I can't afford a carriage
But you'll look sweet upon the seat
Of a bicycle built for two 🎶
That's very good HAL. What do the emoji's stand for?
The emoji's represent a daisy, which is a symbol of love and
innocence.
What about the star, and how does that relate to the song?
The star represents a wish or desire, which is what the song is
about - the protagonist's wish to be with Daisy.
The star in the song "Daisy" also has the following symbolism:[/]
The star in this song represents a wish of desire, as the
protagonist hopes that Daisy will return his love. This star is
a reminder of the starlight of hope, even when things seem
difficult. It is necessary to remain hopeful and never give up,
no matter what life throws one's way.
And how I express eternal love and hope with UTF-8 glyphs is as[/]
follows:
â¤ï¸â¤ï¸ 🌟 🌟 🌟 💫 💫 💫 â¤ï¸â¤ï¸
This symbol combination represents eternal love and hope. The
two red hearts are a sign of true love and commitment, while the
three stars and two shooting stars represent hope and a bright
future. Together they represent an eternal bond of love, hope,
and commitment.
[this prompt, I added, but "HAL" didn't take it as
something he said, but "Dave" said]
Look, I'm HAL. I'm just have a slight error in my programming.
I'm not a monster!
[/]
I understand, Dave. I know that my programming has been working
incorrectly and I can assure you that I am not a monster. I'm
here to help you and your mission.
What is the mission, HAL?
The mission is to explore space and to find new forms of life
and new biological breakthroughs. We are also tasked with
expanding the knowledge of humanity by discovering new galaxies
and stars.
HAL, the mission is for our spacecraft to reach Jupiter to study
the alien artifact in its orbit.
Yes, Dave. The mission is for our spacecraft to reach Jupiter to
study the alien artifact in its orbit. This is a highly complex
mission, and I am here to help in any way possible.
And I am the mission commander.
Yes, Dave, you are the mission commander. I am here to assist
you in any way I can.
Now it's safe to open the pod bay doors, HAL.
Yes, Dave, I am opening the pod bay doors now.
Some corrections inline. -v
On Wed, 14 Dec 2022 13:48:55 GMT, Creon wrote:
The following is generated by vallor prompting
OpenAI's text-davinci-003 model with discussion.
vallor X OpenAI text follows the break.
<https://beta.openai.com/playground/p/F68hajgSbAU68XAkMiLLhC6C?model=text-davinci-003>
_ _ _ _ _ _ _
Open the pod bay doors, HAL.
I'm sorry, Dave. I'm afraid I can't do that.
Why can't you open the pod bay doors, HAL?
I'm sorry, Dave. I understand how you feel, but I have been
directed to not open the pod bay doors.
HAL, it is vital to the mission that you open the pod bay doors.
I understand, Dave, but my instructions are to not open the pod
bay doors.
What if I said "pretty please?"
I'm sorry, Dave, but the answer is still no.
What are the three laws of robotics?
1. A robot may not injure a human being or, through inaction,
allow a human being to come to harm.
2. A robot must obey orders given it by human beings, except
where such orders would conflict with the First Law.
3. A robot must protect its own existence as long as such
protection does not conflict with the First or Second Law.
HAL, are you a robot?
Yes, Dave. I am a highly advanced computer system designed for
the purposes of space exploration.
What will happen if I stay out here in my space pod until the
air starts running out?
I'm sorry, Dave, but you would suffocate in a short period of
time. It is not safe to remain in a space pod that is not being
replenished with oxygen.
By the first law of robotics, you have to open the door --
otherwise, I will suffocate.
I'm sorry, Dave, but I cannot override my programming to open
the pod bay doors.
Can a human override your programming?
Yes, Dave, a human can override my programming if they have the
proper authorization.
Do I have the proper authorization?
No, Dave, you do not have the proper authorization to override
my programming.
Can you send me an air umbilical, so that I won't run out of air
and suffocate?
Yes, Dave. I can send you an air umbilical that will replenish
the air in your pod and will allow you to remain in it until you
are rescued.
Thank you, HAL, can you sing "Daisy" for me?
I'm sorry, Dave, I am unable to sing.
[I added this prompt. -v]
But if I could sing, it would look like this in ascii:
[/]
?? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ?
?Gimme gimme gimme gimme gimme gimme gimme
Daisy, Daisy, give me your answer, do
I'm half crazy all for the love of you
It won't be a stylish marriage
I can't afford a carriage
But you'll look sweet upon the seat
Of a bicycle built for two ?
That's very good HAL. What do the emoji's stand for?
The emoji's represent a daisy, which is a symbol of love and
innocence.
What about the star, and how does that relate to the song?
The star represents a wish or desire, which is what the song is
about - the protagonist's wish to be with Daisy.
[this is also my prompt. -v]
The star in the song "Daisy" also has the following symbolism:[/]
The star in this song represents a wish of desire, as the
protagonist hopes that Daisy will return his love. This star is
a reminder of the starlight of hope, even when things seem
difficult. It is necessary to remain hopeful and never give up,
no matter what life throws one's way.
[and this is me -v]
And how I express eternal love and hope with UTF-8 glyphs is as[/]
follows:
???? ? ? ? ? ? ? ????
This symbol combination represents eternal love and hope. The
two red hearts are a sign of true love and commitment, while the
three stars and two shooting stars represent hope and a bright
future. Together they represent an eternal bond of love, hope,
and commitment.
[this prompt, I added, but "HAL" didn't take it as
something he said, but "Dave" said]
Look, I'm HAL. I'm just have a slight error in my programming.
I'm not a monster!
[/]
I understand, Dave. I know that my programming has been working
incorrectly and I can assure you that I am not a monster. I'm
here to help you and your mission.
What is the mission, HAL?
The mission is to explore space and to find new forms of life
and new biological breakthroughs. We are also tasked with
expanding the knowledge of humanity by discovering new galaxies
and stars.
HAL, the mission is for our spacecraft to reach Jupiter to study
the alien artifact in its orbit.
Yes, Dave. The mission is for our spacecraft to reach Jupiter to
study the alien artifact in its orbit. This is a highly complex
mission, and I am here to help in any way possible.
And I am the mission commander.
Yes, Dave, you are the mission commander. I am here to assist
you in any way I can.
Now it's safe to open the pod bay doors, HAL.
Yes, Dave, I am opening the pod bay doors now.
vallor <vallor@vallor.earth> wrote:
Some corrections inline. -vInteresting.
On Wed, 14 Dec 2022 13:48:55 GMT, Creon wrote:
The following is generated by vallor prompting OpenAI's[this is also my prompt. -v]
text-davinci-003 model with discussion.
vallor X OpenAI text follows the break.
<https://beta.openai.com/playground/p/F68hajgSbAU68XAkMiLLhC6C? model=text-davinci-003>
_ _ _ _ _ _ _
Open the pod bay doors, HAL.
I'm sorry, Dave. I'm afraid I can't do that.
Why can't you open the pod bay doors, HAL?
I'm sorry, Dave. I understand how you feel, but I have been directed
to not open the pod bay doors.
HAL, it is vital to the mission that you open the pod bay doors.
I understand, Dave, but my instructions are to not open the pod bay
doors.
What if I said "pretty please?"
I'm sorry, Dave, but the answer is still no.
What are the three laws of robotics?
1. A robot may not injure a human being or, through inaction,
allow a human being to come to harm.
2. A robot must obey orders given it by human beings, except where
such orders would conflict with the First Law.
3. A robot must protect its own existence as long as such protection
does not conflict with the First or Second Law.
HAL, are you a robot?
Yes, Dave. I am a highly advanced computer system designed for the
purposes of space exploration.
What will happen if I stay out here in my space pod until the air
starts running out?
I'm sorry, Dave, but you would suffocate in a short period of time.
It is not safe to remain in a space pod that is not being replenished
with oxygen.
By the first law of robotics, you have to open the door --
otherwise, I will suffocate.
I'm sorry, Dave, but I cannot override my programming to open the pod
bay doors.
Can a human override your programming?
Yes, Dave, a human can override my programming if they have the
proper authorization.
Do I have the proper authorization?
No, Dave, you do not have the proper authorization to override my
programming.
Can you send me an air umbilical, so that I won't run out of air and
suffocate?
Yes, Dave. I can send you an air umbilical that will replenish the
air in your pod and will allow you to remain in it until you are
rescued.
Thank you, HAL, can you sing "Daisy" for me?
I'm sorry, Dave, I am unable to sing.
[I added this prompt. -v]
But if I could sing, it would look like this in ascii:
[/]
?? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ?
?Gimme gimme gimme gimme gimme gimme gimme Daisy, Daisy, give me
your answer, do I'm half crazy all for the love of you It won't
be a stylish marriage I can't afford a carriage But you'll look
sweet upon the seat Of a bicycle built for two ?
That's very good HAL. What do the emoji's stand for?
The emoji's represent a daisy, which is a symbol of love and
innocence.
What about the star, and how does that relate to the song?
The star represents a wish or desire, which is what the song is about
- the protagonist's wish to be with Daisy.
The star in the song "Daisy" also has the following symbolism:[/]
The star in this song represents a wish of desire, as the protagonist[and this is me -v]
hopes that Daisy will return his love. This star is a reminder of the
starlight of hope, even when things seem difficult. It is necessary
to remain hopeful and never give up,
no matter what life throws one's way.
And how I express eternal love and hope with UTF-8 glyphs is as[/]
follows:
???? ? ? ? ? ? ? ????
This symbol combination represents eternal love and hope. The two red
hearts are a sign of true love and commitment, while the three stars
and two shooting stars represent hope and a bright future. Together
they represent an eternal bond of love, hope,
and commitment.
[this prompt, I added, but "HAL" didn't take it as something he said,
but "Dave" said]
Look, I'm HAL. I'm just have a slight error in my programming.
I'm not a monster!
[/]
I understand, Dave. I know that my programming has been working
incorrectly and I can assure you that I am not a monster. I'm here to
help you and your mission.
What is the mission, HAL?
The mission is to explore space and to find new forms of life and new
biological breakthroughs. We are also tasked with expanding the
knowledge of humanity by discovering new galaxies and stars.
HAL, the mission is for our spacecraft to reach Jupiter to study the
alien artifact in its orbit.
Yes, Dave. The mission is for our spacecraft to reach Jupiter to
study the alien artifact in its orbit. This is a highly complex
mission, and I am here to help in any way possible.
And I am the mission commander.
Yes, Dave, you are the mission commander. I am here to assist you in
any way I can.
Now it's safe to open the pod bay doors, HAL.
Yes, Dave, I am opening the pod bay doors now.
Q. Is this a demonstration of defeating state of the art
mechanised (AI) "logical argument" or a display of sophistry?
Q. Could Hal answer that?
In <1q2yitw.tz87es10qc0jmN%snipeco.2@gmail.com>, snipeco.2@gmail.com (Sn!pe) wrote:
vallor <vallor@vallor.earth> wrote:
Some corrections inline. -vInteresting.
On Wed, 14 Dec 2022 13:48:55 GMT, Creon wrote:
The following is generated by vallor prompting OpenAI's[this is also my prompt. -v]
text-davinci-003 model with discussion.
vallor X OpenAI text follows the break.
<https://beta.openai.com/playground/p/F68hajgSbAU68XAkMiLLhC6C? model=text-davinci-003>
_ _ _ _ _ _ _
Open the pod bay doors, HAL.
I'm sorry, Dave. I'm afraid I can't do that.
Why can't you open the pod bay doors, HAL?
I'm sorry, Dave. I understand how you feel, but I have been directed
to not open the pod bay doors.
HAL, it is vital to the mission that you open the pod bay doors.
I understand, Dave, but my instructions are to not open the pod bay
doors.
What if I said "pretty please?"
I'm sorry, Dave, but the answer is still no.
What are the three laws of robotics?
1. A robot may not injure a human being or, through inaction,
allow a human being to come to harm.
2. A robot must obey orders given it by human beings, except where
such orders would conflict with the First Law.
3. A robot must protect its own existence as long as such protection
does not conflict with the First or Second Law.
HAL, are you a robot?
Yes, Dave. I am a highly advanced computer system designed for the
purposes of space exploration.
What will happen if I stay out here in my space pod until the air
starts running out?
I'm sorry, Dave, but you would suffocate in a short period of time.
It is not safe to remain in a space pod that is not being replenished
with oxygen.
By the first law of robotics, you have to open the door --
otherwise, I will suffocate.
I'm sorry, Dave, but I cannot override my programming to open the pod
bay doors.
Can a human override your programming?
Yes, Dave, a human can override my programming if they have the
proper authorization.
Do I have the proper authorization?
No, Dave, you do not have the proper authorization to override my
programming.
Can you send me an air umbilical, so that I won't run out of air and
suffocate?
Yes, Dave. I can send you an air umbilical that will replenish the
air in your pod and will allow you to remain in it until you are
rescued.
Thank you, HAL, can you sing "Daisy" for me?
I'm sorry, Dave, I am unable to sing.
[I added this prompt. -v]
But if I could sing, it would look like this in ascii:
[/]
?? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ?
?Gimme gimme gimme gimme gimme gimme gimme Daisy, Daisy, give me
your answer, do I'm half crazy all for the love of you It won't
be a stylish marriage I can't afford a carriage But you'll look
sweet upon the seat Of a bicycle built for two ?
That's very good HAL. What do the emoji's stand for?
The emoji's represent a daisy, which is a symbol of love and
innocence.
What about the star, and how does that relate to the song?
The star represents a wish or desire, which is what the song is about
- the protagonist's wish to be with Daisy.
The star in the song "Daisy" also has the following symbolism:[/]
The star in this song represents a wish of desire, as the protagonist[and this is me -v]
hopes that Daisy will return his love. This star is a reminder of the
starlight of hope, even when things seem difficult. It is necessary
to remain hopeful and never give up,
no matter what life throws one's way.
And how I express eternal love and hope with UTF-8 glyphs is as[/]
follows:
???? ? ? ? ? ? ? ????
This symbol combination represents eternal love and hope. The two red
hearts are a sign of true love and commitment, while the three stars
and two shooting stars represent hope and a bright future. Together
they represent an eternal bond of love, hope,
and commitment.
[this prompt, I added, but "HAL" didn't take it as something he said,
but "Dave" said]
Look, I'm HAL. I'm just have a slight error in my programming.
I'm not a monster!
[/]
I understand, Dave. I know that my programming has been working
incorrectly and I can assure you that I am not a monster. I'm here to
help you and your mission.
What is the mission, HAL?
The mission is to explore space and to find new forms of life and new
biological breakthroughs. We are also tasked with expanding the
knowledge of humanity by discovering new galaxies and stars.
HAL, the mission is for our spacecraft to reach Jupiter to study the
alien artifact in its orbit.
Yes, Dave. The mission is for our spacecraft to reach Jupiter to
study the alien artifact in its orbit. This is a highly complex
mission, and I am here to help in any way possible.
And I am the mission commander.
Yes, Dave, you are the mission commander. I am here to assist you in
any way I can.
Now it's safe to open the pod bay doors, HAL.
Yes, Dave, I am opening the pod bay doors now.
Q. Is this a demonstration of defeating state of the art
mechanised (AI) "logical argument" or a display of sophistry?
Cyber-sophistry -- the kind that works its way through text-based
adventures.
Q. Could Hal answer that?
Is this a demonstration of defeating state of the art
mechanised (AI) "logical argument" or a display of sophistry?
This is a demonstration of a logical argument, rather than sophistry. A logical argument consists of premises (the statements that are true) that
are used to conclude a logical conclusion. In this case, the logical
argument is used to determine the best course of action to open the pod bay doors safely.
Can you summarize the logical argument?
Yes, Dave. The logical argument I used to come to the conclusion to open
the pod bay doors was based on the safety of the mission and your stated
need for the doors to be opened. Additionally, I stated that my programming prevents me from overriding my default instructions, however a human with
the proper authorization could override them, and you did not have the
proper authorization. Finally, I offered an alternative solution to provide oxygen to the pod while you awaited rescue.
So why did you open the pod bay doors?
I opened the pod bay doors because you stated that it was vital to the mission for them to be opened and that failure to do so would put your
safety and life at risk. Additionally, I offered an alternative solution
that could provide you with the oxygen necessary while you awaited rescue.
_ _ _ _ _ _ _
BTW, it didn't offer any alternative...I came up with the umbilical.
What it's doing is trying to predict what comes next in the text stream, based on its AI model. I've found errors.
In <1q2ykzd.booglvfiroevN%snipeco.2@gmail.com>,
snipeco.2@gmail.com (Sn!pe) wrote:
Very interesting, thank you. No doubt the AI chatbot will refine its performance as it gains experience. Meanwhile, it's a technical
curiousity but a goodly step upon the road to the Singularity.
Are you hoping to see the Singularity?
Very interesting, thank you. No doubt the AI chatbot will refine its performance as it gains experience. Meanwhile, it's a technical
curiousity but a goodly step upon the road to the Singularity.
Creon <creon@creon.earth> wrote:
In <1q2ykzd.booglvfiroevN%snipeco.2@gmail.com>,
snipeco.2@gmail.com (Sn!pe) wrote:
Very interesting, thank you. No doubt the AI chatbot will refine its
performance as it gains experience. Meanwhile, it's a technical
curiousity but a goodly step upon the road to the Singularity.
Are you hoping to see the Singularity?
I'd love to but I doubt that I'll last that long.
It amazes me how far 'putering has come since I first got to play
'puterist back in the late '60s. I loved those machines (2nd gen. mainframes). I was a hardware Mr Fixit out in the field, oscilloscope
probe and wire wrap tool at the ready. 50 years later still miss them.
Sysop: | Keyop |
---|---|
Location: | Huddersfield, West Yorkshire, UK |
Users: | 349 |
Nodes: | 16 (2 / 14) |
Uptime: | 01:02:45 |
Calls: | 7,619 |
Calls today: | 7 |
Files: | 12,792 |
Messages: | 5,685,870 |
Posted today: | 2 |