Skip to content

Commit

Permalink
Suha's talk info.
Browse files Browse the repository at this point in the history
  • Loading branch information
lkuper committed Nov 12, 2023
1 parent f61b9b2 commit 61f4645
Showing 1 changed file with 27 additions and 6 deletions.
33 changes: 27 additions & 6 deletions content/lsd-seminar/2023fa.md
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ Talks will be advertised on the [ucsc-lsd-seminar-announce](https://groups.googl
| [Oct. 20](#oct-20) | Adrian Lehmann | VyZX: Formal Verification of a Graphical Language |
| [Oct. 27](#oct-27) | Elaine Li | Multiparty Session Type Projection and Subtyping with Automata |
| [Nov. 3](#nov-3) | Karine Even-Mendoza | GrayC: Greybox Fuzzing of Compilers and Analysers for C |
| [Nov. 17](#nov-17) | Suha S. Hussain | TBD |
| [Nov. 17](#nov-17) | Suha S. Hussain | MLFiles: Using Input-Handling Bugs to Inject Backdoors Into Machine Learning Pipelines |
| [Dec. 1](#dec-1) | Kelly Kaoudis | TBD |
| [Dec. 8](#dec-8) | Susan Tan | TBD |

Expand Down Expand Up @@ -162,11 +162,32 @@ We have implemented our approach in GrayC, an open-source LibFuzzer-based tool.

**Speaker:** Suha S. Hussain

**Title:** TBD

**Abstract:** TBD

**Bio:** TBD
**Title:** MLFiles: Using Input-Handling Bugs to Inject Backdoors Into Machine Learning Pipelines

**Abstract:** The widespread use of machine learning (ML), especially in safety-critical applications,
necessitates robust security measures for ML pipelines. Prior research has demonstrated the
existence of model vulnerabilities, including model backdoors that can compromise the integrity
of ML pipelines. Although many backdoor attacks limit the attack surface to the model, ML
models are not standalone objects. These models are embedded in ML pipelines that involve
multiple interacting components and are built using a wide range of ML tools.
In this talk, I will discuss our investigation of input-handling bugs in ML tools as a vector for
injecting backdoors into ML pipelines. Input-handling bugs are central to the field of
language-theoretic security (LangSec), which advocates for the treatment of inputs as a formal
language in order to develop precise, minimalist input-handling code. Drawing from a LangSec
taxonomy of input-handling bugs, we systematically identified and exploited vulnerabilities with
ML model serialization in popular tools. This process enabled us to construct ML backdoors,
substantiating our claim. In the process, we engineered malicious artifacts, including polyglot
and ambiguous files, using ML model files; contributed to the fickling library; and formulated a
series of guidelines to provide actionable steps to ameliorate this issue. Our investigation brings
to light the risks posed by input-handling bugs in tools to the overall security of ML pipelines,
arguing for an approach that concurrently addresses software security issues in tools and model
vulnerabilities.

**Bio:** Suha S. Hussain is a security engineer on the machine learning assurance team at Trail of Bits.
She has worked on projects such as the safetensors security audit and fickling. She received
her BS in Computer Science (with threads in people and theory) from Georgia Tech where she
also conducted research at the Institute for Information Security and Privacy. She previously
worked at the NYU Center for Cybersecurity and Vengo Labs.

# Dec. 1

Expand Down

0 comments on commit 61f4645

Please sign in to comment.