-
We at RiskIQ are authorized representatives of JPMorgan Chase. It has come to our attention that the content located at the following URLs contains accidental data leakage that contains JPMC code/scri…
-
We have observed in your paper that " We use the ShareGPT 4V [2] dataset as a training dataset, which contains about 1M (long caption, image) pairs. The random 1k data is separated as an evaluation da…
-
http://www.idappcom.com/db/?9616
-
### Search first
- [X] I searched and no similar issues were found
### What Happened?
When using `Export public pages` feature, exported files include a number of private information about the grap…
-
Thank you for your great work! But I'm a bit confused. When using SDv1.4 for training, may it is unfair to use all other unseen types of generated images as the validation set. To be fair, only SDv1.4…
-
https://github.com/microsoft/dstoolkit-hierarchical-multilabel-classification/blob/093a988bfb3a0d4c4711d5fe9bec9ce645cbd8e3/src/hmlc.py#L488
In the prep_input function, you use scalers (standard, p…
-
https://arxiv.org/abs/2307.01881
This article provides some promts that researchers have used to test for the possibility of information leakage.
According to the paper, the authors used the fo…
-
The RSITMD and RSICD datasets have a data leakage issue where they might share some common images and descriptions. how to deal with it properly?
-
### Issue Type
Documentation Feature Request
Add a pitfall warning in the docs of "tf.data.Dataset", where the `shuffle` method using together with `reshuffle_each_iteration=True` (which is **T…
-
It would prevent data leakage if it is possible to add padding to encrypted values. Especially when encrypting "enum" values. A good use case for padding is shown on the [Lockbox README](https://githu…