Open prayas7102 opened 3 weeks ago
Hi, following our discussion at #15 I would like to tackle this. Could you please elaborate on
Organizing data and cleaning data.
Hi, following our discussion at #15 I would like to tackle this. Could you please elaborate on
Organizing data and cleaning data.
organizing data: i was thinking if we can combine all csv data into one for training. (let me know your opinion):
cleaning data: as you can see in csv datasets there are rows in which the code contains letters, like this: \n
, //
, empty lines. example, see row no.s like 24, 29 in bruteForceDataset.csv
While utilizing the Naive Bayes classifier to detect brute force attacks, validate inputs, identify insecure authentication, and analyze security headers, the model currently trains separately for each JavaScript test file (.js, .jsx, .tsx, etc.). This process can be streamlined by generating a single (or multiple depending upon the vulnerability) weighted pickle model, which can be reused each time a JavaScript file is tested for vulnerabilities, improving efficiency and consistency.
Steps to be considered by the contributor:
Files to be referred/altered for this change:
Make sure the end user/developer (who downloads the NPM package) is able to smoothly run the NPM package after these changes.