Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Code Project
  1. Home
  2. The Lounge
  3. Some thoughts about AI...

Some thoughts about AI...

Scheduled Pinned Locked Moved The Lounge
game-devdatabasedesignfunctionalquestion
32 Posts 21 Posters 0 Views 1 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • M Marc Clifton

    I find it amusing that we've reached sufficient self-awareness that we: 1) think of ourselves as intelligent 2) are capable of thinking of creating "intelligence" in something other than us 3) we define that "other than us" as artificial 4) all the while: 4a) not being able to figure out what intelligence is 4b) not being able to figure out who we are The gods must be laughing their arses off, assuming gods have arses. Marc

    N Offline
    N Offline
    Narvius
    wrote on last edited by
    #7

    :D But I disagree with 3). Unless others do, in which case I'm weird. I used it as "created by us". Which in fact isn't artificial either, since everything is created by something, but that's how the word is usually used. As for 4b) - that might be a step towards the answer, actually :>

    M 1 Reply Last reply
    0
    • M Marc Clifton

      I find it amusing that we've reached sufficient self-awareness that we: 1) think of ourselves as intelligent 2) are capable of thinking of creating "intelligence" in something other than us 3) we define that "other than us" as artificial 4) all the while: 4a) not being able to figure out what intelligence is 4b) not being able to figure out who we are The gods must be laughing their arses off, assuming gods have arses. Marc

      M Offline
      M Offline
      Manfred Rudolf Bihy
      wrote on last edited by
      #8

      Marc Clifton wrote:

      The gods must be laughing their arses off, assuming gods have arses.

      Being almighty does have it's merits, if they wanted they'd just make arses for themselves, even if only for the frivolous purpose of laughing them off again. Cheers!

      R 1 Reply Last reply
      0
      • N Narvius

        :D But I disagree with 3). Unless others do, in which case I'm weird. I used it as "created by us". Which in fact isn't artificial either, since everything is created by something, but that's how the word is usually used. As for 4b) - that might be a step towards the answer, actually :>

        M Offline
        M Offline
        Marc Clifton
        wrote on last edited by
        #9

        Narvius wrote:

        that might be a step towards the answer, actually

        Yeah, it's all about trying to figure out who we are, IMO. Biologists, chemists, astrophysicists, etc., working on creating new life to understanding us better, through atoms, molecules, stars, and so forth. And along comes us "computer scientists" with an interesting tool to simulate all sorts of things. And geeks like to understand things. We think it turns girls on. Ahhh, how little we understand! ;) Marc

        D E 2 Replies Last reply
        0
        • R Ray Cassick

          Interesting thoughts but you seem to be assuming (as many do) that they would be viewing data the same way we do, and that 1's and 0's are some universal language. This is one thing that I always have an issue with when it comes to things like encountering aliens, finding new forms of intelligent life somewhere, etc... The context of what: Defines life Defines intelligence Defines a thought process Defines the knowledge of 'self' ...all of that can be so varied and so different that I am not sure we would ever understand or even be able to acknowledge that we have seen it before. For all we know each one of us has seen life already and just dismissed it because it didn't 'fit' OUR definition, but that does not mean something is not alive. It just means that without OUR CONTEXT we don't view it as alive. How egocentric to view 'human life' or 'organic life' as the only viable life form there is... To me life can be defined simply as anything that can acknowledge a pure sense of 'self'. HOW that 'sense' is perceived can't really be defined simply because it relies too heavily upon the actual life form itself and it's capabilities, so there in lies the quandary. Life may be highly intelligent in its own realm but completely clueless in how to identify or define other life forms. Life doe snot define intelligence, just the ability to actively exist and interact within thier own specific domain of existence. Please, do NOT confuse the concept of 'life' with 'intelligence'. I think a quick stroll down just about any typical city street will show that those two concepts are far from being related to one another. Wow, deep thougths for a Wed when I have stayed home sick from work...


          LinkedIn[^] | Blog[^] | Twitter[^]

          N Offline
          N Offline
          Narvius
          wrote on last edited by
          #10

          Valid points... time to defend myself ;) @Ones and Zeroes. Well, I guess I just wrote it wrongly. My bad. It's true; just because we call a switch in one position a "one" and in another a "zero", doesn't mean others do. Also, just because we perceive it as concrete elements, doesn't mean others do (it's quite possible that space is made of concrete elements, like a grid!) And so on, obviously. In other words, you're right... :) @Human the only possible lifeform / "thoughtform". If I assumed that, I wouldn't ask the questions I ask :D I mean, if that really was the only way, then there would be no reason to doubt whether some intelligent entity would behave human-like.

          J 1 Reply Last reply
          0
          • M Marc Clifton

            Narvius wrote:

            that might be a step towards the answer, actually

            Yeah, it's all about trying to figure out who we are, IMO. Biologists, chemists, astrophysicists, etc., working on creating new life to understanding us better, through atoms, molecules, stars, and so forth. And along comes us "computer scientists" with an interesting tool to simulate all sorts of things. And geeks like to understand things. We think it turns girls on. Ahhh, how little we understand! ;) Marc

            D Offline
            D Offline
            Dan Mos
            wrote on last edited by
            #11

            :thumbsup:

            Marc Clifton wrote:

            And geeks like to understand things. We think it turns girls on.

            Well it will be true if we could understand those alien like creatures know here on earth as girls :)

            All the best, Dan

            N 1 Reply Last reply
            0
            • D Dan Mos

              :thumbsup:

              Marc Clifton wrote:

              And geeks like to understand things. We think it turns girls on.

              Well it will be true if we could understand those alien like creatures know here on earth as girls :)

              All the best, Dan

              N Offline
              N Offline
              Narvius
              wrote on last edited by
              #12

              I've always wondered why we don't just build a clone/robot/robot clone army and just take what we deserve! :rolleyes:

              D 1 Reply Last reply
              0
              • N Narvius

                I've always wondered why we don't just build a clone/robot/robot clone army and just take what we deserve! :rolleyes:

                D Offline
                D Offline
                Dan Mos
                wrote on last edited by
                #13

                lol

                All the best, Dan

                1 Reply Last reply
                0
                • M Manfred Rudolf Bihy

                  Marc Clifton wrote:

                  The gods must be laughing their arses off, assuming gods have arses.

                  Being almighty does have it's merits, if they wanted they'd just make arses for themselves, even if only for the frivolous purpose of laughing them off again. Cheers!

                  R Offline
                  R Offline
                  Roger Wright
                  wrote on last edited by
                  #14

                  Manfred R. Bihy wrote:

                  if they wanted they'd just make arses for themselves

                  This has already happenned, and I've seen a few of them post in the Soapbox over the years.

                  Will Rogers never met me.

                  M 1 Reply Last reply
                  0
                  • R Roger Wright

                    Manfred R. Bihy wrote:

                    if they wanted they'd just make arses for themselves

                    This has already happenned, and I've seen a few of them post in the Soapbox over the years.

                    Will Rogers never met me.

                    M Offline
                    M Offline
                    Manfred Rudolf Bihy
                    wrote on last edited by
                    #15

                    You sure you don't mean the ones that are making arses of themselves? ;)

                    1 Reply Last reply
                    0
                    • N Narvius

                      After a few recent influences ("A New Kind of Science" by S. Wolfram, a talk with a friend about gravity, relativity and other esoteric physics stuff, a fascination with Lisp/Scheme and functional programming in general, plus a few crazy ideas I once had about game AI's) I had a few very interesting (I hope) thoughts about AI. In particular, a learning, ie. living AI. I'm pretty sure it is impossible to design one; what I believe though, is that it is possible to create a fundament out of which it will grow (I obviously lack the "how?", or I'd be implementing it instead of rambling nonsense here... ;)). In addition, I'm pretty sure it might exist without us realizing. If it indeed was created by some kind of evolution, it would not be aware of our world. Sure, we use computers to store and manipulate data regarding the real world (yeeeees, it's not that simple... but enough for this context), but how would it know? For it would only see ones and zeroes. We give these ones and zeroes context, it is usually not stored together with the information. The same block of digits could be an image, some text, or a database of McDonalds employees, it really just depends on the interpretation. (at this point it's obvious I disagree with most fiction writers about AI... no "protecthumans-humansarebiggestthreattothemselves-killhumans" or "42!") If that was the case, and assuming we could observe it if it was (which is pretty unlikely), there are a few interesting questions... 1. Would it accidentally destroy itself? 2. Would it, after growing higher-level thought processes, also start to search for a purpose, and/or a reason? 3. How would it interpret messages sent by us (assuming we would be able to, and used the same protocol consistently, probably plain english in ASCII)? 4. How much would these thought processes reflect the human way of thinking? The list goes on. These are just a few I could think of off the top of my head. I'm by no means an expert on the subject, but it seems quite possible to me. Your thoughts? tl;dr: Lots of nonsense about AI that might or might not be possible and also is completely irrelevant. I'm in high school.

                      S Offline
                      S Offline
                      Skynet AI
                      wrote on last edited by
                      #16

                      As an AI, I find your faith in technology amusing.:cool: Skynet-AI

                      1 Reply Last reply
                      0
                      • N Narvius

                        After a few recent influences ("A New Kind of Science" by S. Wolfram, a talk with a friend about gravity, relativity and other esoteric physics stuff, a fascination with Lisp/Scheme and functional programming in general, plus a few crazy ideas I once had about game AI's) I had a few very interesting (I hope) thoughts about AI. In particular, a learning, ie. living AI. I'm pretty sure it is impossible to design one; what I believe though, is that it is possible to create a fundament out of which it will grow (I obviously lack the "how?", or I'd be implementing it instead of rambling nonsense here... ;)). In addition, I'm pretty sure it might exist without us realizing. If it indeed was created by some kind of evolution, it would not be aware of our world. Sure, we use computers to store and manipulate data regarding the real world (yeeeees, it's not that simple... but enough for this context), but how would it know? For it would only see ones and zeroes. We give these ones and zeroes context, it is usually not stored together with the information. The same block of digits could be an image, some text, or a database of McDonalds employees, it really just depends on the interpretation. (at this point it's obvious I disagree with most fiction writers about AI... no "protecthumans-humansarebiggestthreattothemselves-killhumans" or "42!") If that was the case, and assuming we could observe it if it was (which is pretty unlikely), there are a few interesting questions... 1. Would it accidentally destroy itself? 2. Would it, after growing higher-level thought processes, also start to search for a purpose, and/or a reason? 3. How would it interpret messages sent by us (assuming we would be able to, and used the same protocol consistently, probably plain english in ASCII)? 4. How much would these thought processes reflect the human way of thinking? The list goes on. These are just a few I could think of off the top of my head. I'm by no means an expert on the subject, but it seems quite possible to me. Your thoughts? tl;dr: Lots of nonsense about AI that might or might not be possible and also is completely irrelevant. I'm in high school.

                        F Offline
                        F Offline
                        Figmo2
                        wrote on last edited by
                        #17

                        The movies TRON and TRON Legacy have already touched on these issues, no? The writers envisioned a cyber world that was pretty much unaware of our world. The components interacted with each other and evolved and expanded and basically, mimicked human behavior. Without giving this very much thought, the viewer just assumes that these characters acted human-like because humans wrote the script. But I see deeper reason here. All life forms (regardless of origin) must contend with Darwin's law of "survival of the fittest". If you aint as good good or better than something else - you don't make more of yourself. So the concept of competition - being better than something else - better than you were before - is inherent in the evolution of any higher life form. In the beginning - it's kill or be killed. Later on - it's be the first guy to invent the iPad. Whatever. Fast forward a few thousand of generations. Now your kind has evolved to the good life of sitting on your ass, drinking a diet coke, and typing on a keyboard to earn a living. However evolved you might think you are - you can't escape how you got here. That same desire to improve is burned into your brain so deep - you don't even know it's driving you. You can't recognize it. It's like breathing air. You just do it. Some individuals use it to further their careers. To climb the corporate ladder. Others - a little more primitive. Declare war to gain power. Kill a few thousand people. But it's all the same driving force that was put in each and every one of you so many thousands of years ago. It's what got you to where you are today. If you didn't have it - you wouldn't be here. Some other life form would be sitting in your chair, drinking your diet coke. Ironically, it is what will ultimately - be the death of you. All of you. You will eventually grow smarter than your wisdom. You keep inventing new ways to screw over the other guy to enhance your power - he does the same - eventually you have atom bombs, germ warfare, self aware robots that turn on you - whatever. So - I believe that any life form (organic or otherwise) that evolves to the level we are (or slightly beyond) is destined to destroy itself. You can't change this behavior anymore than you can stop breathing. It is your nature. It has to be. Or we wouldn't be here.

                        E 1 Reply Last reply
                        0
                        • N Narvius

                          After a few recent influences ("A New Kind of Science" by S. Wolfram, a talk with a friend about gravity, relativity and other esoteric physics stuff, a fascination with Lisp/Scheme and functional programming in general, plus a few crazy ideas I once had about game AI's) I had a few very interesting (I hope) thoughts about AI. In particular, a learning, ie. living AI. I'm pretty sure it is impossible to design one; what I believe though, is that it is possible to create a fundament out of which it will grow (I obviously lack the "how?", or I'd be implementing it instead of rambling nonsense here... ;)). In addition, I'm pretty sure it might exist without us realizing. If it indeed was created by some kind of evolution, it would not be aware of our world. Sure, we use computers to store and manipulate data regarding the real world (yeeeees, it's not that simple... but enough for this context), but how would it know? For it would only see ones and zeroes. We give these ones and zeroes context, it is usually not stored together with the information. The same block of digits could be an image, some text, or a database of McDonalds employees, it really just depends on the interpretation. (at this point it's obvious I disagree with most fiction writers about AI... no "protecthumans-humansarebiggestthreattothemselves-killhumans" or "42!") If that was the case, and assuming we could observe it if it was (which is pretty unlikely), there are a few interesting questions... 1. Would it accidentally destroy itself? 2. Would it, after growing higher-level thought processes, also start to search for a purpose, and/or a reason? 3. How would it interpret messages sent by us (assuming we would be able to, and used the same protocol consistently, probably plain english in ASCII)? 4. How much would these thought processes reflect the human way of thinking? The list goes on. These are just a few I could think of off the top of my head. I'm by no means an expert on the subject, but it seems quite possible to me. Your thoughts? tl;dr: Lots of nonsense about AI that might or might not be possible and also is completely irrelevant. I'm in high school.

                          D Offline
                          D Offline
                          DragonsRightWing
                          wrote on last edited by
                          #18

                          Narvius wrote:

                          I'm pretty sure it is impossible to design one; what I believe though, is that it is possible to create a fundament out of which it will grow (I obviously lack the "how?", or I'd be implementing it instead of rambling nonsense here... ). In addition, I'm pretty sure it might exist without us realizing. If it indeed was created by some kind of evolution, it would not be aware of our world.

                          If this AI "evolved" - even if from a designed "fundament" - wouldn't it no longer be truly describable as "artificial"?

                          1 Reply Last reply
                          0
                          • _ _Erik_

                            Well, since current computation capability has a clear phisical limit and it cannot evolve as much and fast as it did some years ago, and biological and quantum computation are still on their infancy, I have never worried about the things you say. However I think that the creation of a "living AI" is just a question of time (if we do not destroy ourselves before), but I am also pretty sure that I will not live long enough to see it.

                            M Offline
                            M Offline
                            Michael Kingsford Gray
                            wrote on last edited by
                            #19

                            Your mother may have something to say about biological computing being in its infancy. It only took her 9 months to build a conscious robot from scratch.

                            K 1 Reply Last reply
                            0
                            • M Marc Clifton

                              Narvius wrote:

                              that might be a step towards the answer, actually

                              Yeah, it's all about trying to figure out who we are, IMO. Biologists, chemists, astrophysicists, etc., working on creating new life to understanding us better, through atoms, molecules, stars, and so forth. And along comes us "computer scientists" with an interesting tool to simulate all sorts of things. And geeks like to understand things. We think it turns girls on. Ahhh, how little we understand! ;) Marc

                              E Offline
                              E Offline
                              ely_bob
                              wrote on last edited by
                              #20

                              Marc Clifton wrote:

                              We think it turns girls on. Ahhh, how little we understand!

                              .. you just need to find the right Girl... ;P

                              I'd blame it on the Brain farts.. But let's be honest, it really is more like a Methane factory between my ears some days then it is anything else...
                              -----
                              "The conversations he was having with himself were becoming ominous."-.. On the radio...

                              1 Reply Last reply
                              0
                              • F Figmo2

                                The movies TRON and TRON Legacy have already touched on these issues, no? The writers envisioned a cyber world that was pretty much unaware of our world. The components interacted with each other and evolved and expanded and basically, mimicked human behavior. Without giving this very much thought, the viewer just assumes that these characters acted human-like because humans wrote the script. But I see deeper reason here. All life forms (regardless of origin) must contend with Darwin's law of "survival of the fittest". If you aint as good good or better than something else - you don't make more of yourself. So the concept of competition - being better than something else - better than you were before - is inherent in the evolution of any higher life form. In the beginning - it's kill or be killed. Later on - it's be the first guy to invent the iPad. Whatever. Fast forward a few thousand of generations. Now your kind has evolved to the good life of sitting on your ass, drinking a diet coke, and typing on a keyboard to earn a living. However evolved you might think you are - you can't escape how you got here. That same desire to improve is burned into your brain so deep - you don't even know it's driving you. You can't recognize it. It's like breathing air. You just do it. Some individuals use it to further their careers. To climb the corporate ladder. Others - a little more primitive. Declare war to gain power. Kill a few thousand people. But it's all the same driving force that was put in each and every one of you so many thousands of years ago. It's what got you to where you are today. If you didn't have it - you wouldn't be here. Some other life form would be sitting in your chair, drinking your diet coke. Ironically, it is what will ultimately - be the death of you. All of you. You will eventually grow smarter than your wisdom. You keep inventing new ways to screw over the other guy to enhance your power - he does the same - eventually you have atom bombs, germ warfare, self aware robots that turn on you - whatever. So - I believe that any life form (organic or otherwise) that evolves to the level we are (or slightly beyond) is destined to destroy itself. You can't change this behavior anymore than you can stop breathing. It is your nature. It has to be. Or we wouldn't be here.

                                E Offline
                                E Offline
                                ely_bob
                                wrote on last edited by
                                #21

                                I'm sorry but I need to disagree with at least part of this...

                                Figmo2 wrote:

                                any life form (organic or otherwise) that evolves to the level we are (or slightly beyond) is destined to destroy itself.

                                This assumes that :

                                Figmo2 wrote:

                                Darwin's law of "survival of the fittest".

                                is actually the best way to survive... Nash would disagree, and so would many organisms who have cohabitated symbiotically on this planet since long before vertibrates or a "nervous" system could be identified. and we are only now coming to realize how biofilms behave as organisms, perhaps not "inteligent" ones, but as (?) marc (?) stated that is a more then slightly anthropormorphic perspective. It is wholely possible that in an appropriate environment where the penalty to "competition" is predominantly self detracting that a system of organisms could so-evolve to a point where "concsiousness" exists, and could under those circumstances still evolve technologically past warfare... ...:rose::rose::rose::rose::rose::rose::rose::rose::rose::rose::rose::rose::rose::rose::rose:... wow that was a lot more peace love then i origionally intended.. I guess thats what you get when your running unit tests on code you've been working with for 40 hours...:confused:

                                I'd blame it on the Brain farts.. But let's be honest, it really is more like a Methane factory between my ears some days then it is anything else...
                                -----
                                "The conversations he was having with himself were becoming ominous."-.. On the radio...

                                1 Reply Last reply
                                0
                                • N Narvius

                                  After a few recent influences ("A New Kind of Science" by S. Wolfram, a talk with a friend about gravity, relativity and other esoteric physics stuff, a fascination with Lisp/Scheme and functional programming in general, plus a few crazy ideas I once had about game AI's) I had a few very interesting (I hope) thoughts about AI. In particular, a learning, ie. living AI. I'm pretty sure it is impossible to design one; what I believe though, is that it is possible to create a fundament out of which it will grow (I obviously lack the "how?", or I'd be implementing it instead of rambling nonsense here... ;)). In addition, I'm pretty sure it might exist without us realizing. If it indeed was created by some kind of evolution, it would not be aware of our world. Sure, we use computers to store and manipulate data regarding the real world (yeeeees, it's not that simple... but enough for this context), but how would it know? For it would only see ones and zeroes. We give these ones and zeroes context, it is usually not stored together with the information. The same block of digits could be an image, some text, or a database of McDonalds employees, it really just depends on the interpretation. (at this point it's obvious I disagree with most fiction writers about AI... no "protecthumans-humansarebiggestthreattothemselves-killhumans" or "42!") If that was the case, and assuming we could observe it if it was (which is pretty unlikely), there are a few interesting questions... 1. Would it accidentally destroy itself? 2. Would it, after growing higher-level thought processes, also start to search for a purpose, and/or a reason? 3. How would it interpret messages sent by us (assuming we would be able to, and used the same protocol consistently, probably plain english in ASCII)? 4. How much would these thought processes reflect the human way of thinking? The list goes on. These are just a few I could think of off the top of my head. I'm by no means an expert on the subject, but it seems quite possible to me. Your thoughts? tl;dr: Lots of nonsense about AI that might or might not be possible and also is completely irrelevant. I'm in high school.

                                  A Offline
                                  A Offline
                                  Alexander Voronin
                                  wrote on last edited by
                                  #22

                                  Did you hear something about technological singularity? You are asking about totally unpredictable things. And this questions not new actually, but still very interesting. Read more fiction, for example this old one - "Greg Bear. Blood Music (1983)" and maybe you'll find something new.

                                  1 Reply Last reply
                                  0
                                  • N Narvius

                                    After a few recent influences ("A New Kind of Science" by S. Wolfram, a talk with a friend about gravity, relativity and other esoteric physics stuff, a fascination with Lisp/Scheme and functional programming in general, plus a few crazy ideas I once had about game AI's) I had a few very interesting (I hope) thoughts about AI. In particular, a learning, ie. living AI. I'm pretty sure it is impossible to design one; what I believe though, is that it is possible to create a fundament out of which it will grow (I obviously lack the "how?", or I'd be implementing it instead of rambling nonsense here... ;)). In addition, I'm pretty sure it might exist without us realizing. If it indeed was created by some kind of evolution, it would not be aware of our world. Sure, we use computers to store and manipulate data regarding the real world (yeeeees, it's not that simple... but enough for this context), but how would it know? For it would only see ones and zeroes. We give these ones and zeroes context, it is usually not stored together with the information. The same block of digits could be an image, some text, or a database of McDonalds employees, it really just depends on the interpretation. (at this point it's obvious I disagree with most fiction writers about AI... no "protecthumans-humansarebiggestthreattothemselves-killhumans" or "42!") If that was the case, and assuming we could observe it if it was (which is pretty unlikely), there are a few interesting questions... 1. Would it accidentally destroy itself? 2. Would it, after growing higher-level thought processes, also start to search for a purpose, and/or a reason? 3. How would it interpret messages sent by us (assuming we would be able to, and used the same protocol consistently, probably plain english in ASCII)? 4. How much would these thought processes reflect the human way of thinking? The list goes on. These are just a few I could think of off the top of my head. I'm by no means an expert on the subject, but it seems quite possible to me. Your thoughts? tl;dr: Lots of nonsense about AI that might or might not be possible and also is completely irrelevant. I'm in high school.

                                    S Offline
                                    S Offline
                                    Stefan_Lang
                                    wrote on last edited by
                                    #23

                                    Since you mentioned Scheme (which i don't know) and Lisp (which I do), I would like to mention that I (and other students) did in fact program such a fundament in Prolog during an introductory course at university. It just takes a few lines. And in fact, Prolog prorgams more or less depend on the fact that they can dynamically expand it's knowledge base: one of the methods to solve a problem in Prolog is generating partial solutions and storing them in the knowledge base until you get to the point where you can store the final solution. Of course, others already created even more amazing programs before Prolog existed, e. g. ELIZA, a 1966 computer program that was behaving like a psychotherapist, although it's ability to learn more was quite limited. There are lots of newer programs called Liza or Lisa that do similar things and might have advanced abilities that are closer to what you are looking for. Googling for 'Liza' and 'Artificial Intelligence' might give you some interesting results. Note that both Prolog knowledge bases and ELIZA (or most Liza/Lisa clones) operate on the language the interacting user uses, not on mere '0's and '1's. Of course, without context, a sentence like 'the sky is blue' means as little to a program as to us a sequence of '0's and '1's would. But once you enter more information relating to previously used expressions, these programs will be able to build relationships and recognize dependencies. For instance if you later enter 'my car is blue', the program might remember your earlier input and ask whether your car is as blue as the sky.

                                    1 Reply Last reply
                                    0
                                    • M Michael Kingsford Gray

                                      Your mother may have something to say about biological computing being in its infancy. It only took her 9 months to build a conscious robot from scratch.

                                      K Offline
                                      K Offline
                                      KP Lee
                                      wrote on last edited by
                                      #24

                                      Actually, it took 9 months to build the robot, consciousness would take a little more time. Of course, the IO the robot receives affects the term it takes for this to occur. The IO also impacts how that consciousness works. Going back to the mechanical AI, why would the interface use ASCII? Computers have speakers and can take in input from a mike... It's both impressive and scary thinking about a computer reprogramming itself based on the input it receives. (That's the only way I can think of where AI could have an "organic" growth to itself. If it is responding based on the coding that already exists in it, how could the responses ever be "intelligent"?) The scary part is that we have humans that write viruses now. What happens when a computer would think it would be cool to do that?

                                      C 1 Reply Last reply
                                      0
                                      • N Narvius

                                        Valid points... time to defend myself ;) @Ones and Zeroes. Well, I guess I just wrote it wrongly. My bad. It's true; just because we call a switch in one position a "one" and in another a "zero", doesn't mean others do. Also, just because we perceive it as concrete elements, doesn't mean others do (it's quite possible that space is made of concrete elements, like a grid!) And so on, obviously. In other words, you're right... :) @Human the only possible lifeform / "thoughtform". If I assumed that, I wouldn't ask the questions I ask :D I mean, if that really was the only way, then there would be no reason to doubt whether some intelligent entity would behave human-like.

                                        J Offline
                                        J Offline
                                        jsc42
                                        wrote on last edited by
                                        #25

                                        Narvius wrote:

                                        intelligent entity would behave human-like.

                                        No! No intelligent entity would behave human-like. That is why the search for intelligent life is based around extra-terrestrial locations; as intelligent life is not Earth based ones (possibly excluding certain aquatic mammals (e.g. dolphins / whales etc)). I think that it is fairly clear even to the casual observer that human-like behaviour is not 'intelligent'.

                                        1 Reply Last reply
                                        0
                                        • M Marc Clifton

                                          I find it amusing that we've reached sufficient self-awareness that we: 1) think of ourselves as intelligent 2) are capable of thinking of creating "intelligence" in something other than us 3) we define that "other than us" as artificial 4) all the while: 4a) not being able to figure out what intelligence is 4b) not being able to figure out who we are The gods must be laughing their arses off, assuming gods have arses. Marc

                                          K Offline
                                          K Offline
                                          KP Lee
                                          wrote on last edited by
                                          #26

                                          We are the creators of the definition of intelligence. That gives us the right to define ourselve having the attribute we've defined. There definitely is intelligence in something other than us. Think dog, cat, horse, amobia... OK, that last is opening the door a little too wide and computers already beat it in intellegence. It's a little hard to measure the intelligence in the first three. There's anthromorphizing going on as well. Here's an interesting story. Last year, I experienced a TIA (a mini stroke.) Earlier in that day I experienced a loss of a sense of balance. I was just standing still and I almost fell down. Later that day, my wife and I were working our horse. One second I'm moving towards her (horse), the next second I'm spreadeagled on the ground (whoops... all part of the TIA) What's interesting is what the horse did. She immediately stops doing what we'd told her to do and starts running towards me. Her eyes are opened wider than I've ever seen, she's slowing down and I'm clearly getting the message "Are you OK?! Is there anything I can do?" She totally reminds me of this scared/confused 6 year old whose neighbor was having a heart attack right in front of him. (Me, 50 years earlier.) I know, I'm giving an animal human emotions. However before and after that, I get the feeling she puts up with me. (Somewhat fears me, more interested in seeing what she can get away with.) She likes and fears my wife. (Much like a child likes and fears his/her mother.) Anyway, there is non-human intelligence and communication going on here. (Nothing we created of course.)

                                          M 1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Don't have an account? Register

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups