FINDING · DETECTION
TrafficMoE achieves 97.65% accuracy and F1-score on the ISCX-Tor2016 dataset, substantially outperforming all baselines including the best pretraining-based competitor FlowletFormer (91.16% F1), by separately modeling protocol headers and encrypted payloads via dual-branch sparse Mixture-of-Experts rather than treating them as a unified byte stream.
From 2026-he-trafficmoe-heterogeneity-aware-mixture — TrafficMoE: Heterogeneity-aware Mixture of Experts for Encrypted Traffic Classification · §IV-B, Table II · 2026 · arXiv preprint
Implications
- Tor's multi-layer obfuscation is insufficient against classifiers that independently exploit header structure and payload entropy — circumvention transports must destroy discriminative features in both modalities, not just randomize the payload.
- Pluggable transports should be evaluated against dual-branch classifiers that score headers and payloads separately, not just unified byte-stream models.
Tags
Extracted by claude-sonnet-4-6 — review before relying.