Check out our demo notebook!
We will share a demo instance where you can fine-tune most open source LLMs (mistral, llama, gpt..) with or without privacy guarantees in just a few lines of code, and see for yourself that memorization can be solved!
In your infrastructure, you will just have to install the Sarus app and use Sarus python SDK to launch FT jobs with privacy guarantees. Sarus leverages the available GPU when needed. Then your safe models can be put into prod just as usual!
Data scientists explore, preprocess data and feed it to LLMs without directly seeing the data. Only high-quality synthetic data and differentially-private stats can be retrieved from the clean room. To do so, data scientists use their usual AI and GenAI tools wrapped in the Sarus python SDK.
Differential Privacy guarantees can be included in the LLM fine-tuning process itself, through just a fit parameter. This ensures that no personal data is embedded in the fine-tuned model, thanks to automated Differentially-Private Stochastic Gradient Descent (DP-SGD).
Build a synthetic patient records generator without compromising patient privacy