|
Slacker007 wrote: The majority of people on earth do not understand this.
Sadly, the majority of people on earth lack the intelligence to understand this. How ironic.
Latest Article - A 4-Stack rPI Cluster with WiFi-Ethernet Bridging
Learning to code with python is like learning to swim with those little arm floaties. It gives you undeserved confidence and will eventually drown you. - DangerBunny
Artificial intelligence is the only remedy for natural stupidity. - CDP1802
|
|
|
|
|
The thing a lot of people forget is that we are machines ourselves. We're (very complex) multi-celled organisms, that at some point got our "singularity" and became self-aware.
At some point, that will happen to the very advanced AI, whether we like/believe it or not. And I have a feeling it's not going to be pretty.
|
|
|
|
|
Computers only do EXACTLY what they are told to do. So, no, there is no threat unless a programmer programs it to make poor choices.
Social Media - A platform that makes it easier for the crazies to find each other.
Everyone is born right handed. Only the strongest overcome it.
Fight for left-handed rights and hand equality.
|
|
|
|
|
But that's not true for neural networks. They aren't programmed, they are trained, and they aren't nearly as deterministic as coded programs. They are working on fuzzy logic the same as we do, and they can make mistakes like we do.
Explorans limites defectum
|
|
|
|
|
Dean Roddey wrote: , they are trained It still comes down to what the programmer has made possible. A computer can never think or reason like a human. It's still if else statements at its simplest.
Social Media - A platform that makes it easier for the crazies to find each other.
Everyone is born right handed. Only the strongest overcome it.
Fight for left-handed rights and hand equality.
|
|
|
|
|
again, most people in the world do not understand the point you just made.
|
|
|
|
|
But it's not. You should bone up on DNNs a bit more. There is zero problem domain knowledge coded into a DNN. It's just a set of level driven nodes just as our brain's neurons are. There can be problem domain aware code around a DNN do other parts of the job, but the DNN is NOT just doing something it was programmed to do.
It doesn't matter if you consider it intelligent or not. The fact is it will take in lots of information and which generate a choice not based on being told what choices to make and not based on any inputs it has ever seen before. And, like a human, it can make mistakes similar to how we make them, not off/on right/wrong mistakes but fuzzy mistakes.
Explorans limites defectum
|
|
|
|
|
Dean Roddey wrote: You should bone up on DNNs a bit more I actually intend to. At the end of the days, it's just 0's and 1's based on what some programmer made possible.
Dean Roddey wrote: but the DNN is NOT just doing something it was programmed to do. I get that. But it CAN'T do anything that the code won't allow.
Social Media - A platform that makes it easier for the crazies to find each other.
Everyone is born right handed. Only the strongest overcome it.
Fight for left-handed rights and hand equality.
|
|
|
|
|
The codes doesn't ALLOW anything. That's sort of the point of DNNs. They aren't programs in the sense that most programs are. They are more like meta-programs. The program is just the pipes through which the data flows. The decisions are not made by those pipes, it's made by how the data flowing through those pipes interact with each other, which is why it can deal with information it's never seen before.
That's a fundamental difference.
Explorans limites defectum
|
|
|
|
|
Dean Roddey wrote: why it can deal with information it's never seen before Because some programmer wrote code to do that. It's just code. It can't think. It's not alive.
Social Media - A platform that makes it easier for the crazies to find each other.
Everyone is born right handed. Only the strongest overcome it.
Fight for left-handed rights and hand equality.
|
|
|
|
|
It doesn't matter if it's alive or really 'thinks' by your or my definition of what that means. The fact is that it can make decisions much more in the way that we do than like a software program does. They aren't anything alike really.
That means it can be used for things that regular software programs cannot hope to do. And those things it can do very well are things that are potentially very dangerous to us, because human nature will insure that we use them thusly.
Explorans limites defectum
|
|
|
|
|
All I hear is a lot of general theory from you. I have to disagree with you.
Social Media - A platform that makes it easier for the crazies to find each other.
Everyone is born right handed. Only the strongest overcome it.
Fight for left-handed rights and hand equality.
|
|
|
|
|
OK, the earth is flat, if that's what you want to hear.
Explorans limites defectum
|
|
|
|
|
That's a little extreme. I disagree with you. That's OK.
Social Media - A platform that makes it easier for the crazies to find each other.
Everyone is born right handed. Only the strongest overcome it.
Fight for left-handed rights and hand equality.
|
|
|
|
|
So, here's a good example of why you are mistaken. I challenge you to write a program that can recognize any picture of a banana with high accuracy. You will find that that is very difficult. And, when you are done, you will have a program that only recognizes bananas. If you need to recognize something else, like stock manipulation patterns, you will have write a different program that will also be very difficult.
DNNs don't have to be changed to do different jobs like that. That's a fundamental difference. The same algorithm can recognize a banana or find patterns in financial transactions or understand written characters or recognize sounds in spoken words, without any changes.
That's because it's not a program of if/elses that you write. It's a program that accepts data, lets that data interfere with itself in ways that creates a pattern that gives a confidence level that the input represents this or that. It's nothing like a bunch of if/else statements making hard coded decisions. Nowhere in there is any code written related to 'is this a banana?' at all.
It doesn't make any difference whether it's 'alive' or 'intelligent' at all in terms of the practical impact that's already having on our lives and the vastly larger impact it will have in the future.
Explorans limites defectum
|
|
|
|
|
Thought I'd interject to say the question of sentience has been a matter of some debate in the philosophy circles i run in, in large part because of AI being on the horizon.
I think reasonable people can disagree, as there are certain grounding assumptions we all have to deal with here in terms of the question of what makes us human, what it even means to think, or engage in say, philosophy?
As for me I'd suggest that anything that is a convincing enough illusion of The Real Thing(TM) (whatever that happens to be) is as good as the real thing for any meaningful intent and purpose.
For example, for all I know, we don't have free will either. It might be possible to develop a way to plot my next thought or move. Maybe I'm a calculation in a simulation. But it doesn't matter. Because I have the illusion of will, and it's a compelling enough illusion that it may as well be (to me) the real thing.
So I'd suggest here, that at a certain threshold, we might accept that a computer "thinks" as any other sentient being might, or even as a human might.
I don't know if that can be done in silicon reasonably, but I'm entertaining a hypothetical here, if you'll humor me that.
When I was growin' up, I was the smartest kid I knew. Maybe that was just because I didn't know that many kids. All I know is now I feel the opposite.
|
|
|
|
|
codewitch honey crisis wrote: Maybe I'm a calculation in a simulation. The Matrix has you...
M.D.V.
If something has a solution... Why do we have to worry about?. If it has no solution... For what reason do we have to worry about?
Help me to understand what I'm saying, and I'll explain it better to you
Rating helpful answers is nice, but saying thanks can be even nicer.
|
|
|
|
|
ZurdoDev wrote: it's just 0's and 1's based on what some programmer made possible. Well, actually real neurons are more analog than digital and emulated neurons should copy that behavior.
I have lived with several Zen masters - all of them were cats.
His last invention was an evil Lasagna. It didn't kill anyone, and it actually tasted pretty good.
|
|
|
|
|
And they do of course. The nodes in a DNN calculate a level, usually something like -1 to 1. If they were binary you'd need gigantic numbers of them to achieve the same thing. Like real neurons where the strength of an electrical signal is enough to trigger a chemical emission across the synapse or not, these calculate a level that sort of represents the same thing.
Ultimately it's closer to interferometry than a traditional 'decision graph' type of program. It doesn't make decisions, it creates patterns, and via training it's known that a given pattern represents a particular confidence in a particular result.
And of course DNNs can become the inputs to other DNNs. So it's not one huge neural network, and you probably wouldn't want that even if you could do it. It can be a hierarchy where many DNNs are reporting likelihoods of many different conditions and those are feeding into higher level DNNs that are trained to recognize patterns in those conditions and confidences.
Explorans limites defectum
|
|
|
|
|
ZurdoDev wrote: It still comes down to what the programmer has made possible. A computer can never think or reason like a human. It's still if else statements at its simplest. No. The programer wrote an emulation of a neural network, no more. Whatever the capabilites of the neural network may be, they are totally separate from the emulation or the hardware. You can argue that the topology of the network is all wrong, the number of neurons to low or that the learning method is not adequate. The emulation is a normal deterministic algorithm and may fall short of your expectations in many ways, but you are mistaken when you carry these properties over to the simulated network.
Just look at the currently best version of a neural network we have up to now. A unique copy of it is right between your ears. These neurons are real living cells which work on biochemical basis, no emulation needed here. In many ways these neurons are similar to little transistors or surpass them, because transistors can't strengthen, weaken or wire up new connections at all. The basic layout of this network has been shaped by the namegiver of the evolutionary algorithm. From then on it was on its own. nobody programmed it, not even the genetic code that was it#s blueprint. The human genome does not encode enough information to contain a fresh OS installation. And nobody trained it. It started to train itself by processing inputs even before you were born.
I have lived with several Zen masters - all of them were cats.
His last invention was an evil Lasagna. It didn't kill anyone, and it actually tasted pretty good.
|
|
|
|
|
CodeWraith wrote: living Keyword.
Social Media - A platform that makes it easier for the crazies to find each other.
Everyone is born right handed. Only the strongest overcome it.
Fight for left-handed rights and hand equality.
|
|
|
|
|
No. Not at all. Algorithms are independent of their implementation. A neuron implements a switching function and thus implements an algorithm. This algorithm could probably as well be implemented with relays, electronic tubes, transistors, logic gates, in software or even mechanical springs and gears.A tiny biochemical cell was just mother nature's choice.
I have lived with several Zen masters - all of them were cats.
His last invention was an evil Lasagna. It didn't kill anyone, and it actually tasted pretty good.
|
|
|
|
|
ZurdoDev wrote: Computers only do EXACTLY what they are told to do. So, no, there is no threat unless a programmer programs it to make poor choices.
Yeah, this isn't true anymore. Neural networks are black boxes. You train them to recognize a pattern, but no one can read a set of neural network weights and say how they do it.
|
|
|
|
|
It's absolutely a threat. Not in and of itself, anymore than a knife is. But it's a huge threat just because of human nature. Anyone here who thinks that our current baby stuff is indicative of what's to come is fooling themselves. You only have to look at the massive progress made over the last decade or so and project that forward at even a non-increasing rate for some time to come to know what it's going to be like. And more likely it will continue to improve quite non-linearly.
Will it really be intelligent? Not really, IMO. But that doesn't matter. It'll be capable of reacting to massive amounts of input, finding patterns very fast, and making decisions. That will make it irresistible to a lot of players who don't have our best interests at heart.
And, despite the fact that there will have been by that time thousands of books and movies (fiction and non-fiction) predicting the bad consequences of putting such AI's (or whatever you want to call them) in charge of dangerous toys or in charge of us, it's going to happen as sure as the sun rises. Even if every government says it's not going to do it, it'll still be done secretly on the assumption that everyone else is doing it secretly. And it'll become an arms race, both in the weapons world and in surveillance (both business and government.)
Everyone will have an 'AI' assistant in their homes which will effectively know everything they do and say and when they do it and say it and to whom. People will happily pay $1000 a pop to install something that no government could ever get away with forcing them to install. And then everyone will immediately start to work hacking them. Massive resources will be (and already are pretty much) used in the correlation of information in uncountable petabytes of data that will be flowing, which will find everything you do on line, as a consumer, on social media, etc... and ultimately in your own home. Everywhere you go you will be recognized by facial recognition systems. We won't drive our cars or fly our airplanes anymore.
Leaving aside weapons systems, most of these things will be happily adopted and paid for by us. Many of the people working on them or financing them will have intentions that are no worse than just a great interest in making them happen (just as with the bomb) to just old fashioned greed. But, it'll all be a huge system of surveillance and control just waiting to be abused.
And they all will be eventually. That will be far, far too juicy a target or tool. Every government and business and criminal organization (where there's a distinction) will be going after these things full on. And the more powerful they become, the worse the consequences when they are compromised and misused. Governments and businesses will be working overtime to create these systems, and other people in government and business and crime will be looking to hack or misuse them. It's always easy to justify misuse of surveillance systems as patriotism, and it's always easy to justify building nastier weapons because you assume the other side is as well.
And of course we (the US) will likely be at the forefront of the development of nasty weapons and surveillance systems, as we always are. And tt doesn't take even a little bit of cynicism to foresee weapons system responses being put into the hands of AIs who can watch for patterns in enormous streams of data (to respond to incredibly fast AI driven attacks on many fronts from the other side.) We are very easily that stupid and paranoid.
Explorans limites defectum
modified 17-May-19 13:17pm.
|
|
|
|
|
Nah! That'll never happen.
*This message sent from my phone AI*
- I would love to change the world, but they won’t give me the source code.
|
|
|
|