Adversarial Machine Learning in the Context of Network Security: Challenges and Solutions

M Khan, L Ghafoor - Journal of Computational Intelligence …, 2024 - thesciencebrigade.com
With the increasing sophistication of cyber threats, the integration of machine learning (ML)
techniques in network security has become imperative for detecting and mitigating evolving …

Learning graph neural networks for image style transfer

Y Jing, Y Mao, Y Yang, Y Zhan, M Song… - … on Computer Vision, 2022 - Springer
State-of-the-art parametric and non-parametric style transfer approaches are prone to either
distorted local style patterns due to global statistics alignment, or unpleasing artifacts …

A survey on non-autoregressive generation for neural machine translation and beyond

Y Xiao, L Wu, J Guo, J Li, M Zhang… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Non-autoregressive (NAR) generation, which is first proposed in neural machine translation
(NMT) to speed up inference, has attracted much attention in both machine learning and …

Redistributing low-frequency words: Making the most of monolingual data in non-autoregressive translation

L Ding, L Wang, S Shi, D Tao, Z Tu - … of the 60th Annual Meeting of …, 2022 - aclanthology.org
Abstract Knowledge distillation (KD) is the preliminary step for training non-autoregressive
translation (NAT) models, which eases the training of NAT models at the cost of losing …

Knowledge graph augmented network towards multiview representation learning for aspect-based sentiment analysis

Q Zhong, L Ding, J Liu, B Du, H Jin… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Aspect-based sentiment analysis (ABSA) is a fine-grained task of sentiment analysis. To
better comprehend long complicated sentences and obtain accurate aspect-specific …

Understanding and improving lexical choice in non-autoregressive translation

L Ding, L Wang, X Liu, DF Wong, D Tao… - arXiv preprint arXiv …, 2020 - arxiv.org
Knowledge distillation (KD) is essential for training non-autoregressive translation (NAT)
models by reducing the complexity of the raw data with an autoregressive teacher model. In …

Quantum Computing and AI in the Cloud

H Padmanaban - Journal of Computational Intelligence and …, 2024 - thesciencebrigade.com
The intersection of quantum computing and artificial intelligence (AI) within the cloud
environment represents a paradigm shift in the capabilities of computational technologies …

Improving neural machine translation by bidirectional training

L Ding, D Wu, D Tao - arXiv preprint arXiv:2109.07780, 2021 - arxiv.org
We present a simple and effective pretraining strategy--bidirectional training (BiT) for neural
machine translation. Specifically, we bidirectionally update the model parameters at the …

SlotRefine: A fast non-autoregressive model for joint intent detection and slot filling

D Wu, L Ding, F Lu, J Xie - arXiv preprint arXiv:2010.02693, 2020 - arxiv.org
Slot filling and intent detection are two main tasks in spoken language understanding (SLU)
system. In this paper, we propose a novel non-autoregressive model named SlotRefine for …

Rejuvenating low-frequency words: Making the most of parallel data in non-autoregressive translation

L Ding, L Wang, X Liu, DF Wong, D Tao… - arXiv preprint arXiv …, 2021 - arxiv.org
Knowledge distillation (KD) is commonly used to construct synthetic data for training non-
autoregressive translation (NAT) models. However, there exists a discrepancy on low …