Latest news about Bitcoin and all cryptocurrencies. Your daily crypto news habit.
Many call artificial intelligence (AI) a âblack boxâ, and it kinda is. One of the biggest problems of AI is that itâs incredibly difficult to understand how the data is being interpreted.
Before we get our hands dirty and dive deeper, letâs play a little game.
Iâm going to show you a series of abstract images that are either in category A or B.
Do you think the following image belongs to category A or B?Hint: Thereâs no C.
Weâll get back to this later.
Letâs look at some more examples first.Now can you tell if it belongs to A or B?
â ïž Spoiler Alert
The answer isâŠÂ A!
If you choose B, donât be embarassed, youâre not alone. When asked to a room full of engineers and developers, the split is always 50/50. So⊠why is the answer AâŠ?
Because I said so.
The answer is A, thereâs no debating it, but if you donât agree with me, then it was my fault as the trainer.
As the trainer, I know that A is a red circle. So anything with a red circle in it is A. I also know that B is an orange circle. The rest of the image is irrelevant. Itâs all about trying to find a pattern between the set of images.
But itâs hard.
In an AI system, I canât explain with words what makes the image A. All I can do is show you more pictures and hope it starts to click.
And you, the AI, canât tell me why you think itâs B. Itâs up to me to blindly feed you data until you get the answer right.
Hereâs the same set of images, but less abstract. If I were to ask you the same question, everyone would know right away that A is an apple and B is an orange. This is almost so easy that many people think itâs a trick question. We all know that the hand and background are all irrelevant information, because weâre humans and grew up learning these things, but for AI itâs not a given. It sees images as more abstract and doesnât know what you want it to focus on.
A Miscommunication
Letâs take a look at another toy scenario that shows how we might accidentally communicate the wrong signals to the AIÂ system.
We have a few samples of oak trees. (Itâs a bit cloudy where IÂ live)
Here are some palm trees. (It was really sunny on the beach)
This next example is a palm tree, but the lighting is much closer to the oak trees. Which pattern should we focus on? The lighting? Or the shape of the tree? It might be difficult for the model to tell.
Confidence: - Palm 0.75 - Oak 0.60
With this example, it might be pretty obvious that we left behind an unintended pattern for the AI to pick up. However, in reality, itâs normally something much more inconspicuous.
Peeking Under the Curtain
So how can we get more insight into what the AI is focussing on?
What if we passed a rectangle over the image and recorded the changes in confidence? If the confidence drops, then thatâs probably an important part of the image.
Which picture makes it easier to tell that this cable is a USB?
The first image completely obscures the connector, making it nearly impossible to guess, so we can denote the region the rectangle covers as important. However, the rectangle in the second image doesnât hinder our ability to determine the cable type. We can safely mark the location as insignificant.
We can continue to pass the rectangle over image to establish a heat map of importance.
We can see that the modelâs focus is on the tip of the connector, which is great. Itâs looking where we want it to.
Letâs look at a model that wasnât trained well.
Confidence: - USB 0.76
The model correctly predicted that the cable was a USB with a confidence of 0.76. We might say thatâs acceptable, especially since the photo is far away and isnât great quality.
However, upon closer inspection, the model seems to be focusing on the wrong area, not the ends of the cable like we would expect.
What does this tell us?
The model appears to rely too heavily on the wire and fingers. To improve accuracy and clear up the confusion, we can include more examples of wire and hands in a negative training set.
We donât need to train on piles and piles of generic data until our model starts performing better. We can tactfully use this information as an aid in retraining the model, saving us time and money.
Using the Tool
Wow! This is great, but I donât want to put in the effort to actually implement this
Good news! You can find the fully functional iOS app on my GitHub đ
Final Thoughts
Creating your own model is easy, but that doesnât mean the work stops there. The hardest part of machine learning is always producing good data.
We can use the basic guidelines of having similar pose, lighting and a consistent mix of stock and natural photos across our training images to gain a foothold in our quest toward a good model. After that, we are left using tools and our intuition to try and gain insight into the thought process of AI.
Thanks for reading! If you have any questions, feel free to reach out at bourdakos1@gmail.com, connect with me on LinkedIn, or follow me on Medium and Twitter.
If you found this article helpful, it would mean a lot if you gave it some applauseđ and shared to help others find it! And feel free to leave a comment below.
Understanding What Artificial Intelligence Actually Sees was originally published in Hacker Noon on Medium, where people are continuing the conversation by highlighting and responding to this story.
Disclaimer
The views and opinions expressed in this article are solely those of the authors and do not reflect the views of Bitcoin Insider. Every investment and trading move involves risk - this is especially true for cryptocurrencies given their volatility. We strongly advise our readers to conduct their own research when making a decision.