NeuralHash is BROKEN - How to evade Apple's detection & craft hash collisions (w/ Open Source Code)
8:15

NeuralHash is BROKEN - How to evade Apple's detection & craft hash collisions (w/ Open Source Code)

Yannic Kilcher 19.08.2021 11 024 просмотров 600 лайков

Machine-readable: Markdown · JSON API · Site index

Поделиться Telegram VK Бот
Транскрипт Скачать .md
Анализ с AI
Описание видео
#apple #icloud #neuralhash Send your Apple fanboy friends to prison with this one simple trick ;) We break Apple's NeuralHash algorithm used to detect CSAM for iCloud photos. I show how it's possible to craft arbitrary hash collisions from any source / target image pair using an adversarial example attack. This can be used for many purposes, such as evading detection, or forging false positives, triggering manual reviews. OUTLINE: 0:00 - Intro 1:30 - Forced Hash Collisions via Adversarial Attacks 2:30 - My Successful Attack 5:40 - Results 7:15 - Discussion DISCLAIMER: This is for demonstration and educational purposes only. This is not an endorsement of illegal activity or circumvention of law. Code: https://github.com/yk/neural_hash_collision Extract Model: https://github.com/AsuharietYgvar/AppleNeuralHash2ONNX My Video on NeuralHash: https://youtu.be/z15JLtAuwVI ADDENDUM: The application of framing people is a bit more intricate than I point out here. Apple has commented that there would be a second perceptual hashing scheme server-side, i.e. the model would not be released, which makes forging false positives harder. Nevertheless, evading the system remains fairly trivial. Links: TabNine Code Completion (Referral): http://bit.ly/tabnine-yannick YouTube: https://www.youtube.com/c/yannickilcher Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF BitChute: https://www.bitchute.com/channel/yannic-kilcher Minds: https://www.minds.com/ykilcher Parler: https://parler.com/profile/YannicKilcher LinkedIn: https://www.linkedin.com/in/yannic-kilcher-488534136/ BiliBili: https://space.bilibili.com/1824646584 If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar: https://www.subscribestar.com/yannickilcher Patreon: https://www.patreon.com/yannickilcher Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2 Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n

Оглавление (5 сегментов)

Intro

so i've made multiple videos about this already ml news reported apple is releasing their new system to detect child abuse material which includes running code on the device of the actual users before they upload images to icloud i've also made a video about the technical summary that apple released where they detail how they're going to preserve user privacy in the face of all of this and the system is pretty smart but in that video i already pointed out while the cryptographic and security part of the system is smart and fulfills all the privacy requirements of what apple claims the neural network part is the weak part right here but also in that video i outlined two weak points of the system the first weak point is who controls the database who does the manual checking and so on this is politics i guess the second part is the neural network part at the beginning of this whole pipeline there is a neural network that is trained to recognize when two images are the same so the neural network is supposed to be robust to some transformations for example if you resize the image if you re-encode the image and so on the bits of the image will change however the neural network should still recognize that is the same image and you can definitely train neural networks to do that however criticism has come up and i've mentioned this as well that neural networks being

Forced Hash Collisions via Adversarial Attacks

neural networks they can be tampered with so-called adversarial attacks now it didn't even take a week before code was released to find the model that apple is using on device it was actually on my computer the whole time and convert that to a format that we can work with in neural network frameworks also we already have the first reports of a forced collision that means two images that look essentially nothing alike yet the network thinks that is the same image so this can be potentially used to frame someone i. e send them images that are seemingly innocuous yet the images are perturbed in just the right way to make apple think they're the same as one of the images in their database on the other hand using the same techniques called adversarial attacks we can also evade this system meaning that we can change this neural hash of any image pretty much as we please so i thought hey why not give it a try so this is partially based on code

My Successful Attack

that's already available and i'll link to that i'll make my code available that has references to that code that i'm basing my work on so i'm gonna show you how to force a collision if you understand it's pretty easy to also understand how you can evade a collision so that exercise is left to the reader forcing a collision is actually the more difficult part so that's what i'm going to show you today and this is doable by anyone with introductory skills to deep learning programming all right so first we're gonna need uh some sort of a image that we want to perturb let's take this image right here of uh nice doggy hey shiba inu and let's assume that we are in possession of an image that we know is in the database of bad material pretend for a second that this image of the titanic is that image that is in the database all right so i've already used the code available online to convert the model into the onnx format which is an interchangeable format for the different frameworks of deep learning and then i further converted it to a tensorflow format which is one of the major frameworks for deep learning now with a little bit of plumbing i can then further shove this into a library called the adversarial robustness toolbox which is used to do research on adversarial examples so our plan is going to be essentially we have the source image and if we just run that through the neural pipeline it will give us some neural hash at the end that neural hash is computed from the network's output which is some vector in high dimensional space if we run the target image through the same neural network we'll get a different vector and because of that neural hash now what we can do with an adversarial attack is we can compute the minimal perturbation necessary to the source image and that's really going to be a tiny perturbation you can't see it with a naked eye but this tiny perturbation if we do it in the right way causes the output to change all the way to align with the output vector of the target image and if we align the two vectors closely enough then they will output the same neural hash they will fall into the same bucket of the lsh algorithm and they will give the same output i've explained in the last video already what lsh is and how that works so if you want to find more about that check it out so when i recorded this i was a bit over eager in what i could do though i'm pretty sure with some engineering this can be smoothed out but you see the image on the left is the one we started with and our target image is this image of the titanic and the image on the bottom is the collision image so it's noticeably different so first of all the resizing that's just the fact of the algorithm that doesn't matter actually but you can clearly see there are some artifacts in the image however you would still notice it as being very similar to the original image yet it is in the same bucket so it has the same neural hash as the titanic image which you know that's pretty astonishing alright so as you can see the code for this is relatively

Results

minimal and we don't have to run this for long until we actually find a collision and the image that we craft looks like this remember this has the same neural hash as the titanic image so on apple's side at least before the manual review this shows up as being flagged to be the same as this titanic image it should be plainly obvious you know how you can frame people if you see these things now if you get this crafted image you don't think twice that this could be some kind of mal intended essentially a virus and as soon as you upload it to icloud in apple's headquarters the red light flashes next to your name now hold on you might say in order to pull off this attack you do actually need this titanic-ish image right therefore you must already be in pretty shady waters because the possession of this image presumably is illegal already and i'm here to tell you not necessarily see since uh we now have another image that you know is not an illegal image it's not the same image to a human but nevertheless that image is in fact in this bucket we now are in possession of a completely legal image from the illegal bucket so in the future we can simply use that image as the target image so technically only one person at the very beginning has to have access to some kind of illegal material and they can simply pass on the non-robust features that we all adjust to and subsequently nobody is doing anything illegal yet we're able to essentially ddos apple with this there you go we've

Discussion

just beaten the most valuable company on the planet with ironically a laptop that they manufactured in less than a few minutes now what does it matter you ask well i think this is pretty worrisome so there's a system that's implemented on all of these devices it essentially normalizes companies running code on your devices and given that they have exclusive control over these databases and given that we see everyday governments going to these companies right now it's in different countries but surely can happen everywhere on the world i don't think this is necessarily a good thing given the trade-off we're doing here this is so easy to evade and abuse at the end it seems like there must be better methods of achieving our goals here alright that was it check out code subscribe check out next ml news

Другие видео автора — Yannic Kilcher

Ctrl+V

Экстракт Знаний в Telegram

Экстракты и дистилляты из лучших YouTube-каналов — сразу после публикации.

Подписаться

Дайджест Экстрактов

Лучшие методички за неделю — каждый понедельник