AI Security

sAIfer Lab studies how to test Artificial Intelligence and make it more secure.
0
Researchers
0
Publications
0
Projects
0
Years Experience

The research on AI Security, carried on at Pra Lab and SmartLab both, emerged after the discovery of vulnerabilities specific to AI systems. In domains such as Cybersecurity, securing AI against these issues is essential to provide reliable and resilient systems. Thus, research in AI security is important for designing technologies that can withstand evolving adversarial tactics. This research area includes the following three main sub-topics.

  • AI Attacks and Defenses: in this subtopic, we study the mechanisms behind evasion and poisoning attacks. The former produce carefully-crafted input manipulations that cause errors in the system. The latter study how to inject samples in the training data to cause a high error or specific malicious behaviors in the model after training.

  • Secure AI for Cybersecurity: our research designs AI-driven cybersecurity solutions to detect threats and vulnerabilities. Applications include the study of AI-based web security solutions to detect attacks coming from malicious users, malware detectors able to block these threats before they are executed, and spam filtering tools used to block undesired and potentially dangerous emails in the users’ electronic mailboxes. 

  • Large Language Models Security: we explore and develop techniques to assess the robustness of Large Language Models, integrating security measures to ensure that these powerful tools can be safely and effectively used.

sAIfer Lab provide tools and solutions to advance research on these topics, including:

  • SecML-Torch, a PyTorch-powered Python library to assess the security evaluation of AI/ML technologies against evasion and poisoning attacks;

  • an extension of the previous library, called SecML Malware, ad-hoc for attacking Windows malware detectors;

  • AttackBench, a benchmark framework for fairly comparing gradient-based attacks and developed to identify the most reliable one to use for robustness verification. 

To share our research results and train researchers on these topics, we released our free online course on ML Security, and we recurrently organize the MLSec Seminars, a series of events in which we invite researchers to talk about innovations and recent advancements in Machine Learning security to involve both academics and industrial researchers.

Research Topics

AI ATTACKS AND DEFENSES
AI ATTACKS AND DEFENSESPRA Lab - SmartLab
SECURE AI FOR CYBERSECURITY
SECURE AI FOR CYBERSECURITYPRA Lab - SmartLab

Active research projects

sAIfer Lab

Quick Links

Contact Us

PRA LAB:
Via Marengo, 3 - 09123, Cagliari - Italy

SMARTLAB:
Via Opera Pia 11A, 16145, Genoa - Italy

Affiliations