opensearch-project / OpenSearch

🔎 Open source distributed and RESTful search engine.
https://opensearch.org/docs/latest/opensearch/index/
Apache License 2.0
9.55k stars 1.75k forks source link

[BUG] Issue in XPackedInts Serde #10582

Open mgodwan opened 11 months ago

mgodwan commented 11 months ago

Describe the bug When using XPackedInts class and populating across nodes, the same structure yields different value on the 2 different nodes.

To Reproduce

Following UT can reproduce the issue:

        XPackedInts.Mutable data = XPackedInts.getMutable(323, 32, PackedInts.COMPACT);
        byte[] b = new byte[100000];
        for (int i = 0; i < 323; i ++) {
            data.set(i, i);
        }
        data.save(new ByteArrayDataOutput(b));
        XPackedInts.Mutable data2 = (XPackedInts.Mutable)XPackedInts.getReader(new ByteArrayDataInput(b));
        for (int i = 0; i < 323; i ++) {
            assert data.get(i) == data2.get(i);
        }

Expected behavior The collection should give the same values after applygin serde operations. This does not seem to be happening with the snippet shared. On looking a little through the code, the issue can be related to endianness difference during serialization and deserialization.

Screenshots See To Reproduce

Host/Environment (please complete the following information):

Additional context These classes are used in cuckoo filters, and can create problem there.

mgodwan commented 11 months ago

@nknize @reta Any thoughts on this? Based on the javadoc over the methods, and their usage in the existing cuckoo filter, the serialization and deserialization methods used here should be compatible, but seems like they're not.

mgodwan commented 11 months ago

I do see that we write bytes, and read long, which may possibly contribute to issues due to endianness expected while reading and writing.

Write: https://github.com/opensearch-project/OpenSearch/blob/main/server/src/main/java/org/apache/lucene/util/packed/XPackedInts.java#L172

Read: https://github.com/opensearch-project/OpenSearch/blob/main/server/src/main/java/org/apache/lucene/util/packed/XPacked64.java#L100

reta commented 11 months ago

@mgodwan this is weird, I haven't seen this class used anywhere, and I am not sure if it even works. I rewrote the same test using Apache Lucene PackedInts (the XPackedInts seems to be cloned from?) and it seems to work as expected.

@Test
    public void test() throws IOException {
        PackedInts.Mutable data = PackedInts.getMutable(323, 32, PackedInts.COMPACT);
        byte[] b = new byte[100000];
        Writer writer = PackedInts.getWriterNoHeader(new ByteArrayDataOutput(b), Format.PACKED, data.size(), data.getBitsPerValue(), 1024);
        for (int i = 0; i < 323; i ++) {
            data.set(i, i);
            writer.add(data.get(i));
        }
        writer.finish();

        ReaderIterator data2 = PackedInts.getReaderIteratorNoHeader(new ByteArrayDataInput(b), Format.PACKED, PackedInts.VERSION_CURRENT, data.size(), data.getBitsPerValue(), 1024);
        for (int i = 0; i < 323; i ++) {
            long v = data2.next();
            assert data.get(i) == v;
        }
    }
mgodwan commented 11 months ago

The only usage I see for this class seems to be within the CuckooFilter which in turn is used for, and can impact Rare Terms Aggregation use cases.

I was able to pass the test mentioned in issue description by wrapping the DataInput while reading through EndiannessReverserUtil.wrapDataInput(in). Looks like we may need something on these lines to fix this.

reta commented 11 months ago

Looks like we may need something on these lines to fix this.

Why not to use Lucene's ones?

mgodwan commented 11 months ago

Why not to use Lucene's ones?

Yes, that should help solve this as it addresses the concerns. I can see that the Lucene class uses readBytes while reading due to which such an issue may not be present.