TestNet was crowded. To my opinion it was a great success.
Yesterday I read a most interesting article called 'Free Online Coaching on software testing'.
I especially liked the picture...
But if we apply this vision on Artificial Intelligence... then the question raises… should an Ai be able to learn from another Ai...?
People need other people. So does that mean that Ai does need another Ai as well?
And what about the other way around?
Can a human learn from AI?
Can an Ai have a "personality"? or should it have?
Ai says: good idea. human says: bad idea
Ai says: bad idea. human says: good idea
Ai says: good idea. human says: good idea
Ai says: bad idea. human says: bad idea
Will it always stay possible to determine how an Ai came up with a certain answer, or... just like with humans… the result is a chemical process in the brains which can hardly be explained to another person like in the song. "counting crows - accidentally in love".
I think we need to get start thinking about when the weights should be more on the Ai side and when on the human side...
Well... since Artificial Intelligence is a hot topic in the software test community...
What is it then exactly that would make someone an excellent Ai tester?
For example: would he(/she) understand what type (think Eneagram) of Ai he(/she) is communicating with?
Or is that just a strange thought too?
The scary part of that, is that the world to my opinion is not to the well needed level of understanding how software testing works, let alone.. know how Ai testing should be done...
Which could result in unpredictable effects.
(For example: companies of certain countries hacking other companies on the other side of the world samelessly… (just because it's childsplay for them))
In other words:
There are no (world) rules or laws for using Ai yet… !?
(I don't think there's anything (or it is very little) in the law book, or amnesty international notes for that matter)
Are we okay with that fact?
Or do we (ai and humans) have some (test)work to do?