adaltas / node-csv

Full featured CSV parser with simple api and tested against large datasets.
https://csv.js.org
MIT License
4.05k stars 267 forks source link

Delimiter Discovery #400

Open cawoodm opened 1 year ago

cawoodm commented 1 year ago

Summary

The library should detect common delimiters , or ;.

Motivation

Many international users will be working with commas or semi-colons on a daily basis depending on who generated their CSV. Currently we need to manually parse the first line and see which delimiter is used.

Alternative

let delimiter = csvString.match(/[,;]/)?.[0];
let data = csv.parse(csvString, {delimiter};

Draft

let data = csv.parse(csvString, {
  detect_delimiters: [',', ';'],
};

Additional context

Since it may be difficult to detect "just any" delimiter the developer can pass an array of common delimiters which they expect.

hadyrashwan commented 10 months ago

Hey @cawoodm, Anyone is working on this one ? I take work on it.

wdavidw commented 10 months ago

While I am not against the idea, I can't say that I fully support the idea. However, if you come up with a clean delimiter_auto option, I'll probably merge it.

hadyrashwan commented 9 months ago

@wdavidw I looked on GitHub to see what other people were doing.

node-csv-string detect function, which basically looks for the first occurrence of one of the delimiters, I guess is fine for most cases.

Another more advanced implementation was the detect-csv determineMost function , which looks at a sample and returns the delimiter with the highest occurancy count.

What do you think ?

wdavidw commented 9 months ago

I would tend to discover the character, like in the second methods, after filtering any already used character in options (eg quotes, row delimiters, ...) and general ascii characters ([a-zA-z0-9]) (including accented characters).

hadyrashwan commented 9 months ago

@wdavidw I created a small proof of concept for the auto_delimiter option.

https://github.com/adaltas/node-csv/compare/master...hadyrashwan:node-csv:patch-1

When running tests, the below happens, not sure why:

Question:

Missing parts:

Appreciate your feedback :)

wdavidw commented 9 months ago

I'll take some time to review later. In the mean time, what do you mean by "We are committing the dist files".

wdavidw commented 9 months ago

A few notes for now:

  1. delimiter_auto and not auto_delimiter, __discoverDelimiterAuto and not __autoDiscoverDelimiter
  2. Disabled by default, default value is false
  3. When normalizing the option, add consistency check, for exemple it cannot equal the values of record_delimiter (all those rules require tests)
  4. Dont convert to string, you shall compare values directly inside the buffer
  5. Write more unit tests but in particular one which write data one byte at a time (see https://github.com/adaltas/node-csv/blob/master/packages/csv-parse/test/api.stream.events.coffee#L53 as an example)
  6. My strategy would be to discover delimiter before any parsing is done, here is how I will start my experiment
    1. Work around the __needMoreData, if delimiter_auto is activated, and only in the first line, you shall allocated a safe buffer size dedicated to discovery
    2. Start discovery (your __autoDiscoverDelimiter function) just after bom handling and before actual parsing (https://github.com/adaltas/node-csv/blob/master/packages/csv-parse/lib/api/index.js#L109)
hadyrashwan commented 9 months ago

I'll take some time to review later. In the mean time, what do you mean by "We are committing the dist files".

When I'm working I always see the build files in the dist folders added to git and not ignored.

Some projects add those build files in the git ignore file.

Just want to make sure that I'm not adding those files by mistake.

NoahCzelusta commented 8 months ago

A couple of comments on the method of detecting delimiters:

Python has a great example of handling these in their implementation. The pandas library uses this implementation but only reads from the first line of the file (here).

carlbleick commented 7 months ago

Are there any plans to open a PR for that? As far as I can see the current changes are only present on a branch.

I would definitely love to see that feature.

vincerubinetti commented 5 months ago

Here's another algorithm for detecting the delimiter that seems like a good idea: https://stackoverflow.com/a/19070276/2180570

wdavidw commented 5 months ago

I didn't have the time yet. The solution needs to deal with the streaming nature of the parser. The solution would be to extract a limited amount of bytes from the stream, apply a detection algorythm such as the one proposed above, then replay the bytes stored on the side with the detected delimiter. I need some time to do it correctly.

schibrikov commented 1 week ago

Also highly interested in that. My current workaround is to clone the original stream, read the first chunk on a duplicate, detect the delimiter, then get back and handle the first stream.

So basically:

const [mainStream, delimiterStream] = stream.tee();
const reader = delimiterStream.getReader();
const { value } = await reader.read();
const delimiter = discoverDelimiter(value);
reader.cancel();

return { delimiter, stream: mainStream };

Looks somewhat messy, but it works. I have to dynamically handle different types of files avoiding static configuration as much as possible.