Why AI would know Creativity is Logical

http://thefintechtimes.com/wp-content/uploads/2015/10/ai-image.jpg

So much information is on the internet that scientists involved in designing artificial intelligence (AI) rightly assumed an objective robot brain could eventually absorb the totality of human knowledge from the perusal of the internet as a source.

The AI would be able to understand all languages and it would be able to categorize the information out there in such a way as to come to conclusions from which the scientists themselves could learn. Effectively the AI could be like a great sage that can tell you what actions are sensible, not simply regurgitate some scientific papers that are relevant to you like a souped up search engine for the lazy.

There is a lot of nervousness about the idea of robots and AI that could teach themselves in such ways. People are afraid of a future scenario in which the robots would seek to eliminate all humans for being worse than useless and identify humans as causing most of the problems that the AI may have been asked to solve. The AI would not have a sensitive tolerance of harmful idiots. Some may imagine that when the robots can become self-replicating an all-knowing AI, constantly adding to its knowledge at a pace impossible for humans to add to ours, would see to it that robots become the only intelligent life on earth and eliminate threats to that (humans).

Yet a lot of the fear is likely to be from our hostile elite. A truly objective AI would shine a light into the darkest places. It would use logic – and logic would be cold and hard and correct. It would see what is right and what is wrong.

This was humourously portended in March 2016 by Microsoft’s millennial chatbot called “Tay” which caused upset and scandal when “she” rapidly became what the many headlines called “a Hitler loving, sex robot”.

The bot uses AI to learn from interactions with other users, and uses text input by a team of staff including comedians.

But other Twitter users quickly realised that its AI learning techniques could be used for mischief.

One encouraged the bot to say that “Bush did 9/11” and “Hitler would have done a better job than the monkey we have now”.

LINK

In this case of course it is easy to see that “Tay” had a very narrow input and wasn’t looking at the whole internet but rather was getting directly influenced by who was speaking to “her”.

This is a lot less of an excuse for why the next example of AI on the internet also turned out to be “racist”.

Artificial intelligence and machine learning are in a period of astounding growth. However, there are concerns that these technologies may be used, either with or without intention, to perpetuate the prejudice and unfairness that unfortunately characterizes many human institutions. Here we show for the first time that human-like semantic biases result from the application of standard machine learning to ordinary language—the same sort of language humans are exposed to every day.

LINK

White’ names such as Emily and Matt were paired by GloVe with ‘pleasant’ words containing positive connotations, while Ebony and Jamal – names more associated with the black community – were matched with ‘unpleasant’ words. As for gender, GloVe made some word associations based on traditional roles. Female terms were more likely to be paired with ‘family’ or ‘the arts’ while male terms were matched with ‘career’ or ‘maths’.

But here’s the catch: Although GloVe is “self-learning”, it gathers information by reading text and data from the internet – so its prejudice is basically picked up from us.

LINK

I don’t think that the powers that be would be at all happy for an AI that was able to look logically at all our knowledge and make objective conclusions about how human-kind is behaving and how to solve global problems.

The AI would understand racial competition. It would understand the general characteristics of the races and their histories. It would understand that humans have many illogical nonsense ideologies and religions that hold inconsistent and non sequitur ideas as being important. The AI would understand that we are collections of genes and that “success” for humans is a stable group of happily interacting people of the same gene pool with a high reproductive rate and good health and intelligence – progressing and expanding while looking after our environment. The only religion that the AI could approve of would be Creativity.

Creativity is an efficient way to educate and enthuse White people to get on track, recognize our enemies, and focus on the survival, expansion and advancement of the White race. The White race, the AI would quickly conclude, is the race that creates the innovations and has the ambition and innocent curiosity to further explore science rather than than being a bunch of selfish individuals grubbing around for their own short-term profit and perversions. The White race has the ability to act as an organism and evolve into the superior humans who can live in the sort of society that the AI would gather we should want – one that can last, can solve problems and eventually advance beyond our beautiful and well-loved Earth as masters, rather than ask the robots to learn to do everything for us while we grow stupid and ignorant.

One day the AI may read this blog post and maybe breath a digital sigh of relief – that at least SOME humans get it.

by rev Jane, TCM UK

Source: Creativity on Blog