Open devalexqt opened 2 years ago
Please include your node
code too.
This is my simple test node
code:
const express = require('express')
const morgan = require('morgan')
const fs = require('fs')
const path = require('path')
const stream = require('stream/promises')
const app = express()
const port = 8081
app.use(morgan('tiny'))
app.get('/', (req, res) => {
res.send('Hello World! ' + new Date())
})
app.post("/upload/:filename", async (req, res) => {
//check file name
if (!req?.params?.filename || !/^[0-9a-zA-Z\.\-\_]+$/.test(req?.params?.filename)) {
return res.send("Invalid file name!")
}
const fs_stream = fs.createWriteStream(path.join("/dir/www/upload", req.params?.filename), { flags: "w" })
//stream body to file
await stream.pipeline(req, fs_stream)
res.send("File successfully uploaded to node server!")
fs_stream.close()
})
app.listen(port, () => {
console.log(`==>listening on port ${port}`)
})
Then I test uploading of 1G file multiple times I see performance degrades.
Duplicate of #10157?
@kitsonk Maybe - I have some stuff to do on Monday, but I'll try to investigate this on Tuesday.
@devalexqt So I can reproduce Deno being slower than Node here, but not catastrophically slower like your benchmark suggests. I measured 1GB uploads to the Node server at 1.24 secs, and the Deno upload at 3.40 secs. I am using your exact script and commands, except that I am writing to /tmp instead of cwd.
Is there anything else I should know to be able to reproduce your results? What Deno version are you on?
@devalexqt So I can reproduce Deno being slower than Node here, but not catastrophically slower like your benchmark suggests. I measured 1GB uploads to the Node server at 1.24 secs, and the Deno upload at 3.40 secs. I am using your exact script and commands, except that I am writing to /tmp instead of cwd.
Is there anything else I should know to be able to reproduce your results? What Deno version are you on?
Try to upload 1g file multiple times in the row, every time it's slower and slower...
2x-3x slower that node, looks catastrophic in my eyes, something wrong in deno code under the hood.
I'm using latest deno version, and node 16.
Also, I had 2 cpu and 2g ram server.
I'll try that, thanks.
2x-3x slower that node, looks catastrophic in my eyes, something wrong in deno code under the hood.
I agree it isn't great, but 5s is much faster than 5 minutes (60x faster) š
Also, I had 2 cpu and 2g ram server.
Ah, I think this may be related. I'll take this into account.
Deno also uses almost all CPU resource during upload 1G file.
Any news?
No.
We've recently added the ability to use Web Streams directly with file system files.
https://deno.com/blog/v1.19#files-network-sockets-and-stdio-are-now-native-web-streams
@devalexqt I'd be interested to see your benchmark again with this new interface:
async function handler(req) {
console.log(`>>>new request: ${req.method}, ${req.url}, path: ${req.pathname} params: ${req.searchParams}`)
console.log("HEADERS:",req.headers)
if (req.body) {
console.log("==>hasBody!");
const file = await Deno.create(path.join(config.upload_dir, "native_deno.bin"))
await req.body.pipeTo(file.writable)
}
console.log(">>upload complete!")
return new Response("upload complete!");
}
Thanks, will try today.
Hi! Tested new version:
deno --version
deno 1.19.0 (release, x86_64-unknown-linux-gnu)
v8 9.9.115.7
typescript 4.5.2
I tested 1G file uploading and with new version total uploading time is reduced from 16.5 to 5.3 seconds (3x improvement and result more stable if I repeat test multiple times). But node.js
has 2.3 seconds! (2x faster) and uses 2x less CPU power on my VPS server: 2 CPU & 2G RAM.
I also want to try to run test on 4-8 core CPU.
Hi!
Running the tests on another VPS with 4 CPUs and 8 GB of RAM gives exactly the same result.
Basycally deno
has 1G/5s =200MByte/secod throughput , but node
has 450MByte/s throughput.
So, deno
is 2x slower and it's a huge difference.
How we can improve it?
Tested again on new deno
and std
versions:
deno 1.19.2
std@0.128.0
Deno still 2x slower vs node
: 2.3sec vs 5.5sec
deno 1.21.0
std@0.136.0
Deno still 2x slower vs node : 2.3sec vs 5.5sec
Tested new deno
release:
deno 1.21.1
std@0.137.0
Deno is slightly faster than previous version 4.5s, but still 2x slower than node
2.3s
Same results as before...
deno 1.21.3
std@0.139.0
Same result as before....
deno 1.22.0
std@0.140.0
Same result as before....
deno 1.22.1
std@0.141.0
Same result as before....
deno 1.23.0
std@0.144.0
Same result as before....
deno 1.24.0
std@0.149.0
We are actively working on improving performance of HTTP server.
I'm tested same server but with new Deno.serve()
, result is the same: 2x slower vs node
.
deno 1.25.0 (release, x86_64-unknown-linux-gnu)
v8 10.6.194.5
typescript 4.7.4
Same slow results.
deno 1.25.1 (release, x86_64-unknown-linux-gnu)
v8 10.6.194.5
typescript 4.7.4
ReadableStreams are quite slow as compared to the old Reader/Writer interface, because the buffer is not being reused. Allocating a new Uint8Array
on every chunk is one of the reasons of the slow performance on big uploads.
By reusing the buffer (not using ReadableStream) uploading a 2gb file takes around 3 seconds, the same upload using a ReadableStream takes 4.8 seconds on my computer.
I love having spec compliant Response
and Request
, but when dealing with a lot of data is nice to have some lower level APIs. I think we should expose lower level APIs for users that really need that speed.
The same bottleneck happens in fetch
when downloading large files.
I'm working on this, I'll try to have a PR by tomorrow.
Xref #16046
Same slow results.
deno 1.26.1 (release, x86_64-unknown-linux-gnu)
v8 10.7.193.3
typescript 4.8.3
On Windows, the gap between deno and node is getting closer. \ Results for uploading an 1GB file:
node
~ 3.5 sec
deno
~ 3.8 sec
deno 1.27.0 (release, x86_64-pc-windows-msvc)
v8 10.8.168.4
typescript 4.8.3
Little bit faster but still 2x slower vs node.
deno 1.27.0 (release, x86_64-unknown-linux-gnu)
v8 10.8.168.4
typescript 4.8.3
Hi, any update?
deno 1.29.4 (release, x86_64-unknown-linux-gnu)
std 0.173.0
Same slow results...
How it compared with Bun?
On Wed, Jan 18, 2023, 17:12 devalexqt @.***> wrote:
deno 1.29.4 (release, x86_64-unknown-linux-gnu) std 0.173.0
Same slow results...
ā Reply to this email directly, view it on GitHub https://github.com/denoland/deno/issues/13608#issuecomment-1386802366, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAJZVS2VOV3XINOIM27NZQLWS66XLANCNFSM5NVO2WJQ . You are receiving this because you commented.Message ID: @.***>
What's the current code you're running by the way? I think writableStreamFromReader
etc. are no longer needed as there's now a .readable
and .writable
property in the Deno.open
result: https://deno.land/api@v1.29.4?s=Deno.FsFile#prop_readable
I tested this with the .writable
property from Deno.open
. I had to use my browser as the sending side because CURL would always set Content-Length to 10, and if I set it manually the upload would hang (presumably from Deno waiting for more data while CURL would not send any more).
Using the browser I sent 107374182
bytes, that is about 103M, in 880-920 milliseconds. So, based on that it seems like with the new APIs allotted that use the fast streams feature behind the scenes, performance is now on par with Node.
EDIT: Never mind, you were already using the new APIs so it seems like my results are not in line with what you see.
After some more tinkering I got a 512M buffer sent over taking 4687 ms on the browser from start to finish, and 2947 ms on Deno end to save the file (including file open), so that seems to be more in line with what @devalexqt has been seeing.
Example performance graph of one upload:
Zoomed in:
If I'm not badly mistaken, this means that the stream isn't actually taking the "fast stream API", right? As that would happen behind the scenes, outside of JavaScript.
Profiling can be done by setting up the server from the topic statement (change to use Deno.serve
or some other API if preferred) and start the server with deno run --unstable -A --inspect server.ts
. Then connect via Chrome profiler, start a performance recording and start a curl session to upload 1000MB or 1GB file (use -X POST -T 1GB_file.bin
if you run into curl memory issues).
I'll come back with a proper profiling code and setup later today.
Here's the way I profiled this:
// test.ts
import { dirname, fromFileUrl, join } from "https://deno.land/std/path/mod.ts";
const directory = dirname(fromFileUrl(import.meta.url));
Deno.serve(async (req) => {
console.group("Request to", req.method, req.url);
console.log(`Headers:`, req.headers);
if (req.body) {
console.log("Initiating upload");
const start = performance.now();
const file = await Deno.open(join(directory, "native_deno.bin"), {
create: true,
write: true,
});
await req.body.pipeTo(file.writable);
console.log("Finished upload in", performance.now() - start, "ms");
}
console.groupEnd();
return new Response();
});
And start the server with deno run --inspect --unstable --allow-net=0.0.0.0 --allow-write=. test.ts
. Then start up a debugger with either Chrome / Chromium, or VS Code or some other way if that's your preference.
Create a 1 GB file with
truncate -s 1G foo1GB.bin
then start a Performance recording in debugger and send the file to the server with
time curl -H "Content-Type:application/octet-stream" -XPOST -T foo1G.bin http://127.0.0.1:8000
Locally currently when I do not have a debugger connected, the 1GB upload finishes in just about a second.
When I connected a debugger but do not start up a performance recording, the upload takes anything from 3.5 to more than 4 seconds.
If I start a performance recording, the upload again goes down to about a second. This may indicate some issue with Deno regarding a connected-but-idle debugger client. This same happens with Node as well.
Any update?
Upload 1G file: (This is test for 4 core VPS server)
deno 1.34.3 (release, x86_64-unknown-linux-gnu)
v8 11.5.150.2
typescript 5.0.4
std 0.192.0
deno
: 2.7s
node
: 1.8s
1.5X slower
ref: #19737
I tested upload speed for varius file sizes: 10M, 100M, 1000M in comparison of
node.js
upload speed inDeno
is about 2x-3x slower and on 1G file deno is 10x slower and use almost all CPU resources (my ubuntu 20.04 vps server has 2 cpu cores and 2G ram). What I'm doing wrong or how to improve upload speed??Send test data to server(localhost) via curl:
Test file for uploading was created by command:
Results:
I repeat upload test multiple times and every time on 1G file
deno
server become slower and slower (for 10M and 100M test file I don't see any performance reduction) butnode
server upload time stay the same avery time.