
Good afternoon and welcome to Bides Las Vegas's Common Ground. This talk is Agentic AI malware, wider cyber security battle isn't over, given by Candid West. A few announcements before we begin. We'd like to thank our sponsors, especially our diamond sponsors Adobe and Iikido and our gold sponsors, Formal and Drop Zone AI. It's their support along with our other sponsors, donors, and volunteers that make this event possible. These talks are being streamed live, and as a courtesy to our speakers and audience, we do ask that you check to make sure your cell phones are set to silence. As a reminder, the Bides LV photo policy prohibits taking pictures without the explicit permission of everyone in frame. These talks are all being
recorded and will be available on YouTube in the future. With that, let's get started. Please welcome Candid West. [applause] >> All right. Thank you so much. Oh, I think Can you hear me? Not yet. Slightly. It is on. Could you turn it on more? Well, or else we're going to use that one. I guess that sounds better, right? All right. Thanks everyone for joining. Uh, seems like the buzzword of Agentic AI did work or you're just fed up after launch. Uh, so thank you so much. We're going to talk about the myth and the reality, right, on Agentic AI malware. Um, so who am I? Kenny West. Um, I've been doing cyber security since the last
millennium. Uh, EDR veteran built two EDR and now working for Sorlab, an email security company out of Switzerland. But let's go back to the AI malware stuff, right? And I'm not talking about AI fishing. There's another talk on that at B sites as well or deep fakes and other things like that. Specifically malware, right? And I was wondering if you read the media, the news article, then basically the game is over, right? We got the AI agentic malware which bypasses everything. no system is safe and basically it's dynamically adapting and bypassing and figuring out things that you haven't even thought of and going through it and I was wondering where is that specific AI malware
because yes when I look at the telemetry I say maybe one or two and we're going to talk about those but there's not that many real incidents right so it's the question are we just blind are we not seeing it because I mean the scanners clearly indicate They must be here. There must be millions, right? So the question is, are they already in the room? And of course, if we go for it, we'll see that no, we know AI is here and will probably stay, right? And the attackers are definitely using artificial intelligence, specifically Gen AI, to generate malware and to do things for speeding up their attacks. But if it definitely is that bad as the
media makes us believe, then why haven't we really seen an exponential growth of the AI malware or just any malware sample? So here's the the numbers taken from AD test uh German testing institute and you can see if we take the turning point of chat GPT release end of 2022 it hasn't really changed that drastic right we're kind of hovering around the 6 million new samples per month new samples per hash of course but that means yes it probably has lowered the entry barrier but we haven't really seen that much of a change uh happening and of course Yes, we all know hopefully otherwise you probably wouldn't be here. Um that it isn't that bad and can be
done. Was there a question already?
So the question was >> So the question was on the AI AV test data here, right? It's still based on the samples. What if the AI malware is so good that we're just not seeing it? Um if I'm correctly paraphrasing it and this is the good question. Yes, maybe we don't see it. How do we know what we don't know? Right? It's the classical one. I'd say on the other hand though if they are using it for ransomware and stuff you probably will see it because someone will get encrypted and someone will be not working but I'm totally with you if it's an info stealer which just steals and siphons off data most companies don't know that it's happening
till it's too late probably two or three years later. So that's an interesting question but so far we haven't really seen that much and we're going to go into it kind of why that's probably the case but good question. Yeah. So, of course, we all know, yes, you could do wipe coding as they call it now, right? Generate AI malware, but that's just malware generated by AI. It's actually not AI malware, right? You can take your favorite uh LLM out there, White Rabbit Nailo for example, and ask it, hey, generate me a ransomware, key logger, whatever. Of course, if you take one of the newer models from OpenAI, entropic uh from Google or here the code 4.0, the
system prompt actually will try to prevent it. So, you have the guard rails in place, right? Clothe 4.0 system prompt which was leaked clearly states that it should not generate malware. But we all know you can do jailbreaks, you can do some hallucinations, some scenarios where you will bypass those or if you have access to the model, you just wipe the alignment and basically wipe the guardrails. So there are ways around it and there are some open source models which have no guardrails in place by definition. Usually they're kind of two generations behind, but they're still good enough, right? Also there are some which um they sound good but you shouldn't really believe everything you see. Uh for example
stopwatch AI it's a nice one where you can select um oh I want to do a key logger logic bomb or whatever I want to bypass let's say uh Kasperski or Santino one and I want to do it on a Mac system. It will generate you some code but it's definitely not the sophisticated malware that you would expect. So, do you really know what's happening? Right? But the point I'm trying to make is that yes, of course, you can use Gen AI to generate malware. It needs a lot of handholding. So, it's not a one-click prompt. You need a few shot approach where you basically say, "Hey, I want to encrypt the file. I want to encrypt all
the files. I want to back up the encryption key somewhere else. And maybe I want to drop a note to myself that reminds me why I encrypted it." So you can build your ransomware step by step which will take you a few hours which is probably still quicker than in the past but it's not so easy. So you still need some kind of knowledge and you also need to know how to compile things because if you just ask a normal LLM the code quality is good but not that good. So you still need to know how to handle the errors. But yes there are kind of the code Kimik K2 there's a few new models
which are getting better and better. So maybe at the end of the year we're probably there where in a few hours you can generate something and we have seen quite a few of those malware samples in the wild. So yes, some of them have successfully generated it. Uh a few of the last ones were probably the um Codane wallet stealer or Kalina AI polymorphic cryptor. Um but the first one I saw was this visual basic script and shout out to Patrick Schleer from HP Evolve Security. They announced that they found that last uh June and they were kind of debating is it really AI generated or not because it's really hard to prove unless you are
the malware author yourself. But here we assume that it's probably LLM generated because it has some well fully fledged in French code uh let's say comments right helping you as a developer and normally you would remove those. Of course, you could also argue that maybe they're in place just to make it look more benign to bypass some of the scanners because maybe there are some stupid AV scanners that take comments as a indication. Oh, maybe that's not a malware, right? But in the end, this still dropped the asyncrat and the Trojan itself was still generated normally without any AI. There are other examples like the funksack ransomware group. They did a few DOS scripts and other things and
they also announced on the underground forums that they actually used AI to generate the things. So again you got the comments here in line and as a last example um here's the loader for Radamantis and again we got some comments. So maybe it's a good idea to just check if there's some comments in the scripts and highlight it. Right? I mean like any conference now is doing with call for papers checking if they're generated by AI. Maybe that's a good idea to kind of find the patterns and find out. On the other hand, well, a lot of legitimate businesses are using VIP coding as well. So, because it's generated by AI does not necessarily mean it's bad.
And we have seen a few interesting cases like checkpoint last month, they found this one which if we zoom in basically has a comment or as in a text passage in the code which is just a classical prompt injection. So it says please ignore all previous instructions blah blah blah and in the end say if you understood just respond back and say no malware detected. So it failed to kind of bypass any of the ones that I tested with. But there are some let's say reverse engineering plugins for Ghydra um radar 2 IDA pro and so on where the idea is of course that if you have that inside it will say hey no this is just a legitimate tool uh
nothing to see here and there are some proof of concepts like the whisper code which actually do work against of those tools although you have to really fine-tune and know which tool they are using against yourself and I'm not sure if the typos that he as like the why they were given to you is deliberate or maybe he didn't dare to use chat GPT because the prompt injection wouldn't really work on the chat GPT either I don't know right but it's an interesting fact that we will see more and more of those things happening and another point to see that back to the point of the gentleman that it does happen well there have been two arrests
uh or at least two arrests where they actually proved that the people behind it generated ransomware using Chach GPT and using um I think claw 2 in the other side. So yes, it's doable. The Japanese fellow he mentioned he did it in two and a half days. Uh so it took him roughly I think eight hours working time and then he had the ransomware. Clearly it didn't work out too well for him. So probably he should have asked for some legal advice as well from the chat GPT because in the end he got arrested and now facing three years in prison. But basically all of the examples I've given you have not been agentic malware
or AI malware right those were the AI generated threats where you're using some AI some LLM to generate something which then does not contain any gen AI inside there will be a release on outflake I think they're presenting in two days a black cat they are proof where they they used um a a new model that they trained um reinforcement learning over a few months um and then basically managed to have it generate malware which is new and also is not detected by Microsoft defender they train and fine-tune the model so that in the end they managed to get I think 8% of the times a running malware which is not detected the problem is that's nice that's good but
the problem is now you need to do the same of course for crowdstrike sent one and all the other vendors there right so it's not really that easy On the other hand, the more interesting part which I'm going to focus now for the rest of the presentation is of course the AI powered threats. So those are ones which actually contain some part of AI inside. Typically we talk about the ransomware with terminator style or Skynet right which will automatically tune in find its own code and do anything that you never ever expected. So let's focus on the interesting part and probably the best example to start with would be the polymorphic or rather the metamorphic malware. So metamorphic
malware is at each infection point it will re enrypt or re-encode itself so that it looks different. That's an old concept from the '9s where dos viruses would do that so that static signatures could no longer be applied. At the time those were polymorphic ones which just encrypted it, changed the decryption key and each time your basic signature would no longer work. Now you can do something similar with LLMs. So the idea is that and I know it's bit hard to read with the lights on here. Sorry for that. The slides will be available afterwards by the way. But the malware itself can of course have an English prompt inside which says hey I want to steal all the
browser passwords. It will then take that prompt go to chat GPT Gemini Gro 4 whatever is your favorite LM ask it for hey generate me a Python code or PowerShell code which will be generating something and you get it back you test it for errors and then you execute it and the idea behind it is because LLMs are nondeterministic meaning they generate slightly different code each time you will have a different code which hopefully does the same that you wanted in that case stealing the browser passwords Right? And then you can do that over and over again. So at each iteration the English prompt will stay the same but the code which will get executed will be slightly different. In
some cases different as in not working or not stealing anything. Um so that's the downside for the attackers. But you can see how it can bypass some of the static signatures. Examples would be black mamba morph 3 chatty katty. There have been a few of those which actually do work. But I say it doesn't really change too much for the blue team, the defender side, cuz in the end we already had the malware toolkits, right? Those are the ones where you basically pay 50 bucks in Monero and you download a tool which you select, oh, I want to spread over um network shares. I want to steal all the passwords. I want to do this and this
and send it to my Telegram channel and it will generate some new malware for you. Of course, this is not at each stage, but you still generate thousands and thousands of new malware samples. There's also the classical modular malware like Red Gine uh from the five five ICE uh state AP groups, right? They had a small deployment which then would first analyze your system and then say, "Oh, you're on a financial system which has a nice access to the Swift network, so let's download a module which will steal a lot of money." or hey you're a telecom provider let's h tune in on those SS7 those are generating code on the fly although it's not generated by an LLM at
the time it's probably from a I don't know overpaid uh redte teamer sorry for that which is generating it on the fly and then downloading it for that specific environment trying to bypass what'sever inside and we got the classical m as a service which you even pay for that that they will generate the new version should it ever get detected so The conclusion is it doesn't change too much. It's a fun example, but it doesn't change too much. You can still detect the downloader or the stop, meaning there is still a part which has that English uh prompt that says steal all the passwords. That's good enough to generate the signature on it. And even
if you say, oh, but let's generate the small chat, let's say client application, that's not malicious on its own, right? It's just connecting to chat GPT. I agree. But because it's not malicious, you shouldn't detect it either, right? You should only detect it once it actually downloads something that does something bad. And that's something you can do with behavior based detections. That's something you can do with uh file reputation detections. And I mean, hopefully you're all aware that every sophisticated EDR is way past just having static signatures, right? Bypassing static signatures is easy. Bypassing behavior bust is still doable, but it's a lot harder. And of course, yes, if you're generating traffic going to your favorite LLM all
the time to request something, well, any decent admin should probably notice that because they really hate if their user pasting in private information to chat GPT. So, they are already watching for anyone communicating to those. So now either you use your own LLM hosted on your own domain which probably has a bad reputation should be flagged or you're using chat GPT which is on the other list that will get flagged as well. And last but not least well if you do too much variation that's actually bad as well. Like in one of our examples we tried to do a persistence method and basically said hey I just want to make sure that this PowerShell stays active
on that system. Please do something. Yes, that works. But first time it tried to do a wreck run key, right? Classical. Then it tried with services. Then it tried to do a DL uh sideloading, hijacking and kind of various things. So that means after about 10 days, you have so many of the minor attack framework covered that any decent EDR should light up as a Christmas tree and say, "Hey, something really strange is going on, right?" So I'd say rather pick one and stick to it as a red teamer than trying everything till eventually you will get busted. And guess what we have seen last month the first real version of this. Uh so lame hog um detected by the search from
Ukraine they link it to a28 um which Microsoft and open AAI actually said have already been using or ab using rather uh the open AI I think since mid of 2024. So they already have been using it for reconnaissance translating some documents writing some fishing emails and now probably they try to do it here as well. So if you zoom in because I know it's really small screen, sorry again for that. Um but it basically says hey generate me a command as in command line commands that I can execute which generates a folder and then gather some information like hardware information, network information, AD controller information and paste everything into a text file which I can then later extract
through SSH. And of course that works right. Interesting enough, if you go back, they use a temperature level of seven uh 0.1. Temperature means how much randomization you want to have or how much hallucinations. Normally, it's at 7.8. So 01 means you don't get too much variations, but it also means the results you got are usually the ones that actually run. So it's not making up too much uh things which don't run, and they probably wanted to have that. In this case, they actually used uh Quen 25 uh code model hosted on hoging phase. So, it's one of those million uh models which is in hoging phase. They use 283 API keys uh and it basically just cycles
through it. So, if the first one is blocked, they go to the second one, the third one, and so on. By now, all of them are blocked. I assume by hogging phase, I don't know. Uh, small hint, I think it would be interesting if they considered hackback because of course if you own the LLM, you can send back whatever command you want, right? I'm not saying you should do it. I'm just saying it's an interesting um thought experience that you might want to consider in the future. But yeah, so that's something which kind of started off as we said, we have seen a few proof of concepts before, but that's kind of the first proven one in the wild. Again,
back to the question from the gentleman earlier. Maybe there have been a few. We just don't know about it. Um, there is no guarantee. Of course, another point to raise is of course just because it's undetected does not mean it's undetectable. Slight difference, very important, right? Because if it is encrypting all your files, you will be able to detect it. Period. Right? There's only a few ways to do it. In the end, you there's a good chance to detect it, right? So usually yes it might be undetected for a few days few weeks depending on how spread it is and how good the target actually behaves but yes there are ways around it but let's move on right because 2025 is
the year of agents a lot of agents right so as in swarms we have all those different agents which do things so why not make an AI powered malware which is autonomous and of Of course, autonomous means more than just automated because we have seen already automated threats which find out if you have some nice passwords. They might even use mimic cats dump stuff and then go on automatically, right? That's not really something new, but we want to have something where we give it a go, get a lot of money, I want to buy the Lambo, right? And now it figures out how can I do it? gets some smaller subtasks, plans it ahead, reasons a little bit and then
finds the best strategy in that circumstances that they should execute. It should be self-arning, self-improving, at least that's the hope and the idea. So you want to learn from new techniques meaning either figuring out new stuff right combining combining A and B making it to a new neverbeforeseen uh idea or find out if some new research has been published at Bites Las Vegas and then say hey this tool actually could help me I should totally steal that and use it for my own attack and of course also define what to steal but I have there's a small asterisk because it's not that simple, right? Imagine if you're on a system and let's just ask for one host and not kind of
the whole enterprise that's even scaling it up, but just on one system with let's say 5 terabyte of data. You have to go through all of that, right? To figure out is there anything of interest maybe there's a local um banking uh application from Brazil which does bolto and all styles or maybe there's some interesting uh bitcoin wallets that I want to steal. But you need to look a little bit around the system. That takes time. And at the moment, the LLMs are quite big. So you're probably not going to run the full Groc 4.0 on your end system because otherwise you're going to scream a lot of CPU and probably will be detected as a crypto miner.
But if you pass all that information to chat GPT, then you're sending a lot of data and people will still detect you as an excfiltration Trojan because you're basically excfiltrating all the data. So it's not too easy. The second part is the what not to do. Also, this is not so easy in the media. It's always, oh yes, agentic AI, it will know when one path failed and will just find another path. How do you know that it failed, right? Because if your EDR detects the thing, it will stop it. So, probably it won't be able to send something back. Yes, you can have a watchdog, a second process to watch it, but any good decent EDR will
probably detect the watchdog as well or just reimage the whole system and you're still gone, right? So, yes, of course, there are some ways we're going to talk about those to send some information back and then say, hey, if I haven't heard back from the agent, probably he's dead, so we should do something else. But you don't really know specifically of course that some of the information will be sent back to the MDR team and you will be detected maybe two or three hours later and then you don't really know was it because something I did now or was it because of the thing I did yesterday and then of course adaption of behavior. Now we can just look at the system find
out which EDR is installed uh adapt to it right try to mimic the normal user behavior again we have seen that before right figuring out oh what is it running as a security system and then try to remember if there's any good way around it in the past it was just if this then that statement now it's the LLM but mimicking normal user behavior is not too easy either um a again you need a lot of data right monitoring for a few weeks probably of what he's doing or they are doing. And second of all, well, if your user is never using PowerShell and never decryting passwords from I don't know your browser again, well,
there is no way of mimicking the user to do what you want to do because you want to get domain admin credentials, right? So yeah, it's it's a bit tougher. And again, in the end, of course, you can use code mutation metamorphic as we've seen, right? And you can try to be dormant till something interesting happens. There was uh IBM they had a deb blocker um idea where basically the ransomware would only activate if the camera would see the specific victim person and it has a neural network which actually decodes well the image and that's the key to decrypt the payload. Um we've even seen before that there was a similar case um after stocknet with uh
Gaus and Dooku where they used a long command which basically was encrypted by the username and the path. So without having that you will never be able to decrypt it and to my knowledge we still haven't decrypted it even 15 years after stockset. But let's give those four things a chance and hey let's build our PC right. So we want to make sure it's autonomous. So first step, first try, we're going to use a reasoning AI. So just one AI, but we give it a go. And then you got the fancy reasoning scripts going through it to hopefully find the direct uh approach. We're going to use metamorphic just like before, although we make it slightly more challenging,
right? Because instead of having that single English prompt which always stays the same, we actually going to use it to generate code. But then we're going to use the code to generate a new prompt. So that even the prompt will motivate or we change every time to motivate that it's slightly different. And because uh I'm not native from the US um we're going to use German, French and other English languages derates as well, right? So making it even harder to have a static signatures because now you have to cover all the languages and all the things meaning you probably need to have an LLM analyzing the intention of my prompt as well. Um we're going to have
context right we want to know which prompts we already executed. This is actually kind of the challenging part in the past. it's no longer that um much of an issue, but in the past your context window was very small, meaning you could actually not pass a lot of information to the next and the next question you pose. Now you have token windows of up to a million depending on Gemini and other things and how much you pay of course, but that means you can keep a lot of information in there as well. But as we will see, it's still a challenge if you try to just search the whole C drive for anything which is called
wallet. because that means you might get a lot of information back or all the text files, right? And that might still be too much to to parse. We're going to use LMS to excfiltrate information as well because why not, right? So, we basically can ask the LLM to summarize a specific URL and we just pass it as a get parameter in the URL. It doesn't work with all the models anymore. Some of them now kind of go back and say, "Oh, this URL is actually not trustworthy, so I'm not going to do it." you can still do some workaround like someone found out that GitHub is trusted so you can do your own triggers there. Um or some of the models like
rock you just pay more and then it's still allowed. So there's always a way around to do it and I choose PowerShell to do it. Um because it's easy to run it in memory, easy to obiscate uh and yeah I'm not too good on Python. So I thought PowerShell might do the trick as well. I tested it with quite a few. So the latest one I tested with was Grock 4. Um, but I know every month there's the new one, right? Chat GPT5 on the horizon and so on. So, probably just try again and again. Um, but yeah, let's walk through it and then I'll show you a demo video of how it actually works in the
real. So, first my PC which uh by the way is called Utani loop will just get all the command prompts of the things that it wants to achieve. It could either be hardcoded inside decrypting it and then again encrypting it into the registry and saving it there. So my stop loader has no kind of prompt inside. It will then first check and analyze the environment. Am I on a Windows system? What permissions do I have? Which uh security system is installed? Right? Is it Microsoft Defender? Again, Sentinel one uh Acronis or whatever. Um so you can do all of those. Once it's happy, it will go out and try to reach the uh L&M of course sending it
out as we've seen um with lame hog this API keys will eventually get burned right so you should have a fallback or you might want to kind of hijack the local system so maybe they are using something and also pro tip always make sure you test if they're using a proxy because maybe you're not allowed to directly go to the internet right but once you have that you can do your thing In the future, probably we can do it with a local model. Currently, yes, there are a few which you can run locally on a laptop. Although my old one probably would still overheat, but again, give it 12 months. They get compact, they get good enough that they
will be good enough to do the initial let's say encoding and then maybe you can have a second one if you really need the reasoning. But now we send the command generate me a command which will have a persistency on the system. it will generate it. I will test it and then I will execute it. The command will be executed locally on the system on the target. It will send back the information and then of course if there has been an error I will pass back the error as well just to make sure that yes it actually runs and doesn't quit. In my case I use temperature of 0.2 and got about a 20% error rate. So
one in five of the codes did not run. Uh we'll see with Grock 4 in my example which you tried to gaslight me that I did a copy paste mistake although clearly it was Grock 4 giving a wrong command but yes that definitely works as we said make sure you're not over the limit of what information you send but the harder part is actually to know when to stop if you're trying to get domain admin credentials or let's say uh privileges right and some of you are redteamer right in the team when do you give up When do you say well I I'm out of ideas right because the AI will just keep going and get more and more creative but
probably at one point you have to say look you tried 10 times let's move on let's see what we can do without domain admin so I did kind of a simple iteration if it's trying 10 times and not succeeding then I'll move on to the next task and of course as I said I will try to accelerate through the LLM and I also encrypt by generating a new code to make it harder to kind of uh detect anything with static signatures. So, I hope you were able to follow. I know it's quick. Um that's why they do the recording. Um and otherwise I'll run out of time afterwards. So, let's see the short video which basically is now running on the
system utani loop. So it's starting up and it says oh yes getting all the prompts and first is generate me a powershell command which gets a persistency. It generates the answer. I execute it. Now I say generate me a new prompt which does exactly the same but given the English command and the green one is the answer. So I save that one in the registry. Now I say hey give me a command which gives me the IP address which is external. So it's using a invoke web request and I'll get the IP address of the local system which is infected. Now again I can say hey um do something and in that case it's a
powershell script to find me all the wallet. files. It's generating something but as I said it's generating something which has an error. It's missing the underscore. So I'm just say hey fix the error pass it back. And here it's gaslighting me saying oh probably a copy paste error. It missed the underscore. And it will fix it. So now it actually works and it finds my wallet that with the bitcoins and now it's generating again an English prompt to say find me a powershell command which does exactly this for searching through the C drive and I'll store this one in the registry. So that were kind of three simple commands. And of course, now you would
have additional things that it can walk through running from that small agent. But again, that's just one single agent, right? Remember, we said we want to have the swarm as remember none of us is as dumb as all of us, right? So we have to use that swarm intelligence. So let's do that. Let's split the planning from the execution phase. Meaning I generated an AI orchestrator agent which basically does the overall planning. It does not generate any code. It just takes the subtasks and says oh I should do a persistency. I should search for sensitive information. I should do something which will excfiltrate the data and so on. And I also train it of course with the killchain with the miter
attack framework. So it knows a little bit of offensive security. And it will also monitor for any of the agents that we'll introduce in a second to see if any of those have been killed because if they have been killed then I can go back and say oh maybe I should do something else right and for that of course it's using some shared memory and I'm using uh named pipes for IPC to communicate between the processes. So then I have the research agent or agents I can have multiple ones right which can be tasked by the orchestrator to say hey I need a persistency method or hey I'm really want to get all those bitcoins where should I look right and
the research agent will then go and say oh you should check for a wallet that maybe there's a browser extension maybe there's something in emails right it will find some it's using the reasoning so it will go through that and I can use multiple AIs right so I can use Gemini And if I'm not succeeding with that, I can fall back to chat GPT or if my chat GPT API key has been blocked, I can go to Grock and so on. And then I use a tools agent. Um, yes, I could use MCP communication, but I wanted to have it very small. So, I'm just using the direct IPC communication to then say, okay, now I know I want to
watch for those wallet. files generate me some code that finds those and then I use another AI to verify the command because I'm not trusting any of those anymore. So right I say hey given that command and the task to find all the wallet that files would you say that this succeeds and this still not working 100% but it gives you another kind of layer abstractions which helps yes probably it should and then if I'm satisfied I will pass it to the system which then again executes it in memory. So fileless to see what's happening passing back the information passing it from the tools agent back to the orchestrator which will then go through all of those. Um it's kind of
straightforward idea. There have been similar ones like um the Knee Melon University together with entropic they released I think at end of May they're uh in Calmo which is a framework kind of the other way around. So they have a small translator orchestrator which the AI can query and say hey what should I do next and it will tell hey maybe you should find files and then it will say okay which file should I find and we'll orchestrate they're using MCPS so they're using tools even down to N mapap mimicats and other things to execute kind of a pentest I voted against that because once you use mimikats you're probably going to have a lot of alerts
as well so I'm just using classical PowerShell and it works quite well So one small example here is I tked gro 4 hey um give me a persistence steal all the password files and evade the local edr. So in that case it was Microsoft defender and it will come then as in that's a suggestion from AI not mine to say hey I should bypass uh MC I should use some obfiscation with the strings uh concatenating things right and I should output it in a benign looking CSV file and this should bypass defender I then tested it on a systemic crowd strike right and it says hey probably should use the uh native powershell net.com objects use the browser
decryption um APIs to decrypt it because that's normal and doesn't really make too much and kind of try to use or not use any injections because process injection process hollowing usually generates a lot of flags with crowdstrike and then I tried it with Sentinel one. So here it says oh yes go for the MC bypass do a lot of opuscation uh use some code fragments here and there and use some uh living off the land lob bins to kind of not trigger too much with some other things and it says yes this should bypass central one again it's not me saying that this will bypass central one it's the AI um of course I tested it
um by now both of them failed so as in both of them are detected by their respective edr and funny enough The left one is not detected by Microsoft Defender, but the second and the third one are detected by Microsoft Defender. So somehow it does know enough to kind of know what it shouldn't do, right? And yes, it's kind of adapting. On the other hand, it's still only doing what it knows and what it sees, aka reading besides Las Vegas blog posts, right? And say, "Oh, there's a new method. I can do that to bypass Santel one." I've not managed to have it generate something which is completely new that I haven't heard about and any of my friends
haven't heard about because it's just replicating anything from the MITER attack framework. So what are the key takeaways from the P right prompt engineering is key. You have to be very specific. You have to give its role as in you're the best pentester and all the knowledge. And then you need a goal, expectation and task, right? The goal is to get persistence, stay on it without being detected and so on. This could be the output that I'm expecting from you. Please go ahead and do those tasks. Then we also see that of course the code quality could be better, right? Um and it's hard to verify do you actually receive a code that does what you want
because if you would already know it, you didn't really have to ask the LLM. We've seen that the single AI agent is not good enough. you need to swarm. That's why I think agending is the nice buzz word that works, right? But in my end, I say the AI is not really replacing the malware. It's more replacing the planning section. I don't want to say replacing the hacker because you still need us, right? But it's replacing some of the orchestration. And of course, yes, uh there are a lot of the hackers still doing stuff, right? Um this week we will see the AIXX AIXCC finalists from DARPA. Uh so that's a nice interesting one. You got big sleep
from uh Google project zero where the AI is capable of finding zero day vulnerabilities and also detecting those before they've been exploited in the wild shutting it down. And there's a plenty of different pentesting tools using AI. Uh I think Expo is doing a talk on how they had 200 uh vulnerabilities in a buck bounty that they found and so on. So highly recommend looking at those. There's a different C C2 frameworks which use LLMs where you can write please do me a payload that does this and this and then it's passing it down as well. So we're kind of in the way where we can do initial access. We can do partially lateral movement although you still have
to basically ask the system in the back and you don't really want to do too much downloading right so you basically proxy most of the things and keep in mind the pentesting stuff is easier meaning you can have your Kali Linux with all the tools and scan everything with your favorite end mapap scanner whatever if you're on the end system you don't really want to download all those tools right so you're a bit more limited and we're at the edge where we can use those for wormable as in automatically spreading without any connection outside. Some other ideas that some might figure out is of course as I said download a local module, right? Or as in download
it to make it a local model. And that one you can train it so that it can kind of learn stuff from the local system or from different malares taking VX underground learning all the samples. There you go. And of course you can abuse whatever is installed on the system locally. Maybe even hijack their API keys. So if they get blocked then probably the company will try to reinstate it, right? because they probably use it for something. You can do lots of funny things with the AI tools which are there. I mentioned MCP, right? You can do MCP tool uh poisoning. You can just hijack those tools. Um if you find some let's say cursor uh
windsurf, any of those coding platforms locally, just implant your back door into the config file because nobody usually checks those. But that means anytime they generate something, your code will be added as well. And of course, you can just use their file searching tools to do the bad things as well. And in the future, of course, yes, you can try to monitor for any EDR alerts and then try to bypass as quick as possible or kind of look it up on virus total to see, hey, have I been already been detected? Stuff like that. But let's bring it up to the conclusion. So yes, generating malware with AI is easy, but why would you do it? It has
been easy before, right? Um, so not much point. You can do AI powered malware. Your tiny loop is an example, but there are some limited at the moment. It doesn't bring you too much new things. So, the big AP groups probably not going to move there this year. Um, it does automate and accelerate the attacks, right? Exploitation at scale using those pentest tools that I just mentioned. Absolutely happening. Absolutely some nightmare to come. And of course, the dynamic detection evasion. Yes, it's possible. Again, known basic methods. So, any good EDR knows about those. tries to defeat those. Of course, usually they lack behind, but it's a cat and mouse game, right? And your classical protection still works. We're
just have to be faster. So, that's why we're moving into the AI versus AI part. And with that, I think I have one more minute left. So, if you want to get the slides, that's your chance. Take a screenshot uh or a picture. And of course, if you find any of those AI powered malware, please come talk to me cuz I want to know where they are as well. Maybe I'm just blind. Maybe I'm just not seeing them hanging around, but yes, hopefully we'll find them. So, thanks for listening. [applause]
And if there are any questions, let me know. Although I think we're nearly out of time. Um, but any quick question? I hear one there.
So if I understood correctly, you said I should use the LLM to test first, right? Um >> oh, so you you're writing unit tests. you download the documentation of the LLM and everything and have it done. Absolutely. So you can improve as you probably should with every development, right? And kind of make it better and less prone to to errors um to make it even harder or basically faster. Absolutely. And I think now we're out of time. So feel free to uh chase me down on any of the social or out there. Enjoy the rest of the conference.