Feb. 12, 2024, 5:43 a.m. | YunDa Tsai Cayon Liow Yin Sheng Siang Shou-De Lin

cs.LG updates on arXiv.org arxiv.org

This paper reveals a data bias issue that can severely affect the performance while conducting a machine learning model for malicious URL detection. We describe how such bias can be identified using interpretable machine learning techniques, and further argue that such biases naturally exist in the real world security data for training a classification model. We then propose a debiased training strategy that can be applied to most deep-learning based models to alleviate the negative effects from the biased features. …

bias biases cs.cr cs.lg data data bias detection generalized issue machine machine learning machine learning model machine learning techniques paper performance security training url world

Research Scholar (Technical Research)

@ Centre for the Governance of AI | Hybrid; Oxford, UK

HPC Engineer (x/f/m) - DACH

@ Meshcapade GmbH | Remote, Germany

Encounter Data Management Professional

@ Humana | Work at Home - Kentucky

Pre-sales Manager (Data, Analytics & AI)

@ Databricks | Stockholm, Sweden

Lecturer / Senior Lecturer - Medical Imaging

@ Central Queensland University | Mackay, QLD, AU

Intern - Research Engineer

@ Plus | Santa Clara, CA