This is not a hot dog: an intuitive view on attacking machine learning models

In a previous post we introduced the field of adversarial machine learning and what it could mean for bringing AI systems into the real world. Now, we'll dig a little deeper into the concept of adversarial examples and how they work.For the purpose of illustrating adversarial examples, we’ll talk about them in the context of … Continue reading This is not a hot dog: an intuitive view on attacking machine learning models

Users ignore your security awareness program? Ditch it!

Yes, getting staff attention for security awareness is hard. It's not that users don’t care. But everybody is fighting for their attention. And after all, the company is investing big money on security measures, so they're probably safe anyhow. Way too often, for each handful of truly enthusiastic users I find, there's also a large community … Continue reading Users ignore your security awareness program? Ditch it!

Apples or avocados? An introduction to adversarial machine learning

A common principle in cybersecurity is to never trust external inputs. It’s the cornerstone of most hacking techniques, as carelessly handled external inputs always introduce the possibility of exploitation. This is equally true for APIs, mobile applications and web applications.

It’s also true for deep neural networks.