Open kirudang opened 3 months ago
Thank you for your interest in our work.
We also welcome you to use the relevant code from markllm (https://github.com/THU-BPM/MarkLLM) for testing and experiments. It integrates the SIR algorithm, which is also our team's work, making it more convenient to compare with other algorithms.
Hi there,
Thank you for your prompt response. They definitely resolve my concerns. For the Tool kits, I will take a look at it soon.
Many thanks for your help.
Hello,
I am trying to produce your watermarked texts for my experiments: Set a prompt of first 200 tokens and let Llama 2 with your SIR watermark generate max new 200 tokens. I have a few questions regarding this:
python train_watermark_model.py --input_path data/embeddings/train_embeddings.txt --output_model model/transform_model_cbert.pth --input_dim 1024
` parser.add_argument('--generate_number', type=int, default=2)
In our SIR, it is mentioned "Without a watermark, the expected score is 0 since the watermark logit mean is 0. When a watermark is present, the score substantially exceeds 0". So should I use 0 as detection threshold? Or do you have any suggestion? Some of my output's z:
"z_score_generated": 0.48147795266575283, "z_score_generated": 0.385262405798759