Closed Sankalp-CS21MTECH12010 closed 1 year ago
@Sankalp-CS21MTECH12010 Please refer to this first: https://github.com/networked-systems-iith/AdaFlow/issues/10 Done:
AdaFlow uses Decision Tree to prioritize flows and XGBoost as an offline ML model.
Key takeaways (First Graph):
Observation: When dtM is high, AdaFlow behaves as a malicious traffic filter and gives very good recall at the expense of low precision.
Key takeaways (Second Graph): As we can see, these results are extensions of first graph. Best results are given by dtM = dtB = 0.8. The lowest FNR is given by dtM=1, dtB = 0.
In Progress:
@Sankalp-CS21MTECH12010 --
@praveenabt
Last 5 (important features)
bin75: Say a flow has 100 packets. Then bin75 = 12 for that flow means that out of 100 packets there are 12 packets whose packet lengths lie between [16x75, 16x76 - 1).
How do we show the benefits of Adaflow over NetBeacon?
@Sankalp-CS21MTECH12010 For facet/deltashaper, explain the key reason of why NetBeacon recall is giving poor results compared to AdaFlow. -- Is it because of missing small flow monitoring?
@praveenabt I do not think it is because of missing small flow monitoring. NetBeacon people have clearly showed in the paper that doing that does not affect the metrics much for P2P, Covert and DDoS detection.
I think it is because of ML Model design. NetBeacon uses sequential model while I have used aggregated model design. I wanted to bring this yesterday's meeting but did not because the discussion was already becoming complicated.
So sequential model consumes more switch memory (in terms of SRAM and TCAM) than the aggregated model I used.
We can show that AdaFlow achieves similar metrics (slightly better) than NB for less number of stages (less SRAM and TCAM memory), which is an advantage. The sequential model requires deploying many ML models (depending on a number of inference points).
Another point is that showing that for any CIC-IDS attacks, we cannot miss small flow monitoring could be another advantage.
Lastly, we can also argue that AdaFlow can selectively filter out specific malicious traffic, which NB might not be able to do.
Observation: NetBecon dp model consumes more memory (multiple models deployed) interms of SRAM and TCAM, compared to our single aggregated model.
Important Features for P2P Fingerprinting:
@praveenabt
Uses 10 stages, 8.1% SRAM and 13.9% TCAM
For DDoS (500 lines of code): Used 8 stages, 5.7% SRAM and 1.5% TCAM In contrast NB reported 9 stages, 11.1% SRAM and 1.85% TCAM
For P2P: Uses 11 stages, 12.8% TCAM and 9.6% SRAM In contrast NB reported 12 stages, 17.29% of SRAM and 31.25% of TCAM.
Todo: Get the accuracy/recall/precision, etc plots and memory usage information for all four datasets for both AdaFlow and NetBeacon.
Combined prototype to handle covert channels, P2P and DDoS (1000 Lines):
This is NB output for P2P: Meter ALU is 27.1%. In contrast AdaFlow only consumed 18.4% ALU for P2P.
So there is a decrease, about 8% decrease for P2P and covert. For DDoS about same.
Although I'll have to think why this decrease for meter ALUs.
@praveenabt For P2P and DDoS, I will directly add the results in the draft paper shared.
Rebuttal Reply: We agree that to make a better case regarding accuracy and generality, AdaFlow should be evaluated on different datasets – P2P App Fingerprinting, Covert Channel, and DDoS detection. We will conduct experiments to evaluate FPR, FNR, malicious flow loss, packet recirculations, and ROC curves and complete these experiments by the camera-ready deadline.
Todo items: