Skip to content

ashwin1596/Capstone

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

36 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Ownership Verification in Deep Neural Networks via Watermarking

Overview

As deep neural networks (DNNs) become central to a wide range of real-world applications, the need to protect these valuable models from unauthorized use and intellectual property (IP) theft is increasingly critical. This project investigates watermarking techniques as a means of ownership verification for convolutional neural networks (CNNs), focusing on preserving model functionality while offering reliable IP protection.

Objective

The primary goal of this project is to evaluate and compare different watermarking methods for CNNs with respect to:

  • Robustness under adversarial attacks (e.g., pruning, knowledge distillation)
  • Stealthiness and undetectability
  • Functional impact and performance trade-offs

Techniques Explored

We implemented and analyzed the following watermarking approaches:

  1. Backdoor Trigger-Based Watermarking

    • Embeds unique input-output behavior using trigger patterns.
    • Robust under pruning; fails under knowledge distillation.
  2. Weight Perturbation

    • Introduces small, structured modifications to model weights.
    • Maintains moderate robustness under pruning; fails under distillation.
  3. Passport-Based Watermarking

    • Requires valid cryptographic credentials (passports) at inference time.
    • Uniquely supports intentional performance degradation when used without a passport.
    • Adds an additional layer of deterrence and control.

Methodology

Our study was conducted in two main phases:

  1. Baseline Model Development

    • A CNN model was trained and evaluated on a standard classification task.
  2. Watermark Embedding & Evaluation

    • Each watermarking method was applied to the baseline model.
    • Models were subjected to attacks including weight pruning and knowledge distillation.
    • Metrics such as F1 score and watermark verification accuracy were used for evaluation.

Results

Poster

Results Summary

Method Pruning Robustness Distillation Robustness Unique Features
Backdoor Trigger ✅ High ❌ Low 100% verification accuracy post-pruning
Weight Perturbation ✅ High ❌ Low ~70% verification accuracy post-pruning
Passport-Based ✅ High ❌ Moderate Controlled performance degradation on fail

These results highlight important trade-offs between security, robustness, and usability when selecting a watermarking method for deployment.

Project Structure

├── src/
   ├── models.py              # CNN model implementations
   ├── base/                  # Base model implementations
   ├── backdoor_wm/           # Backdoor Trigger Scheme implementation
   ├── weight_pt_wm/          # Weight Perturbation Scheme implementation
   ├── sign_enc_wm/           # Passport-based Scheme implementation
   ├── README.md              # Project description
├── environment.yaml          # Python dependencies
├── Report.pdf          # Detailed report
├── Poster.pdf          # Poster

Key Takeaways

  • No one-size-fits-all: Each method has strengths and weaknesses depending on the threat model and deployment scenario.
  • Passport-based watermarking introduces a novel control mechanism that deters unauthorized usage via intentional model degradation.
  • Robust watermarking requires balancing stealthiness, reliability, and performance impact.

References

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •  

Languages