privacy-scaling-explorations / acceleration-program

Accelerate Early Stage Programmable Cryptography Talents
100 stars 7 forks source link

Privacy preserving machine learning using MPC #4

Open tkmct opened 1 year ago

tkmct commented 1 year ago

Open Task RFP for Privacy preserving machine learning inference using MPC

Executive Summary

Project Details

Qualifications

Administrative Details

Additional Information

Submission Details

thogiti commented 1 year ago

I just finished PSE's Summer ZK Fellowship program and I have some previous experience in ML.

I want to work on this task.

In the past I worked on Federated Brain Tumor Segmentation from a privacy enabled ML POV.

NOOMA-42 commented 1 year ago

I just finished PSE's Summer ZK Fellowship program and I have some previous experience in ML.

I want to work on this task.

In the past I worked on Federated Brain Tumor Segmentation from a privacy enabled ML POV.

Hi @thogiti Kindly send out your proposal as issue per the template

mitsu1124 commented 1 year ago

Hey @thogiti , update?

thogiti commented 1 year ago

Hi @mitsu1124. Apologies for delay. I got caught up in some stuff. But I did make some notes after doing some self-studying about this project. I will write them down and put it in a proposal and post it here for your review and feedback in the next one week.

Thank you. Apologies again for a delay.

saurabhchalke commented 8 months ago

Proposal: Privacy-Preserving Machine Learning Inference using MPC

Executive Summary

Project Name: Trustless MPC Inferences for Advanced Machine Learning Models

In this project, we aim to extend the capabilities of privacy-preserving machine learning (PPML) by implementing trustless Multi-Party Computation (MPC) inferences on larger and more complex models like Whisper, GPT-2, Mistral 7B, and Gemma 2B. Building on our experience with smaller models such as ResNet and CISER, we will leverage the Crypten library and explore the newly developed mpz library to demonstrate the effectiveness of MPC in maintaining privacy without compromising model performance.

Project Overview

Our focus is to push the boundaries of PPML using MPC by applying it to advanced machine learning models. By ensuring privacy during the inference phase, we aim to enable secure and confidential utilization of state-of-the-art models in sensitive applications. This will also encrypt the model, protecting against weight leaks and whitebox attacks.

Project Details

Scope of Work

  1. Model Selection: Choose larger and complex models for MPC implementation, such as Whisper, GPT-2, Mistral 7B, and Gemma 2B.
  2. MPC Implementation: Extend our work on trustless MPC inferences using the Crypten library to the selected models.
  3. Library Exploration: Explore the mpz library as an early adopter and integrate it into our MPC implementations. Also, consider other libraries like MPCFormer & EzPC.
  4. Evaluation: Assess the performance and privacy-preserving capabilities of the MPC implementations on larger models.

Milestones

Milestone 1: Model, Library Selection, and Preliminary Setup

Milestone 2: MPC Implementation on Selected Models

Milestone 3: Evaluation and Documentation

Team

Name Email GitHub
Gunit Malik gunitmalik@gmail.com @guni7
Saurabh Chalke saurabhchalke@gmail.com @saurabhchalke

Team Experience

The team has been deeply involved in the zk space for over a year. We have previously built privacy-preserving versions of zk proof delegation based on the zksaas paper, utilizing the packed secret-sharing MPC primitive. The team has prior experience in AI, having worked with computer vision, SVM, language models, and with PyTorch/TensorFlow.

Administrative Details

Current Progress

We have successfully implemented trustless MPC inferences on smaller models like ResNet, MNIST, and CISER using the Crypten library. This experience has laid the foundation for tackling larger and more complex models in this project.

@NOOMA-42