Lasuie: Unifying information extraction with latent adaptive structure-aware generative language model

H Fei, S Wu, J Li, B Li, F Li, L Qin… - Advances in …, 2022 - proceedings.neurips.cc
Universally modeling all typical information extraction tasks (UIE) with one generative
language model (GLM) has revealed great potential by the latest study, where various IE …

Pop music transformer: Beat-based modeling and generation of expressive pop piano compositions

YS Huang, YH Yang - Proceedings of the 28th ACM international …, 2020 - dl.acm.org
A great number of deep learning based models have been recently proposed for automatic
music composition. Among these models, the Transformer stands out as a prominent …

Syncobert: Syntax-guided multi-modal contrastive pre-training for code representation

X Wang, Y Wang, F Mi, P Zhou, Y Wan, X Liu… - arXiv preprint arXiv …, 2021 - arxiv.org
Code representation learning, which aims to encode the semantics of source code into
distributed vectors, plays an important role in recent deep-learning-based models for code …

Code prediction by feeding trees to transformers

S Kim, J Zhao, Y Tian, S Chandra - 2021 IEEE/ACM 43rd …, 2021 - ieeexplore.ieee.org
Code prediction, more specifically autocomplete, has become an essential feature in
modern IDEs. Autocomplete is more effective when the desired next token is at (or close to) …

Multi-hop attention graph neural network

G Wang, R Ying, J Huang, J Leskovec - arXiv preprint arXiv:2009.14332, 2020 - arxiv.org
Self-attention mechanism in graph neural networks (GNNs) led to state-of-the-art
performance on many graph representation learning tasks. Currently, at every layer …

Sentibert: A transferable transformer-based architecture for compositional sentiment semantics

D Yin, T Meng, KW Chang - arXiv preprint arXiv:2005.04114, 2020 - arxiv.org
We propose SentiBERT, a variant of BERT that effectively captures compositional sentiment
semantics. The model incorporates contextualized representation with binary constituency …

Image captioning through image transformer

S He, W Liao, HR Tavakoli, M Yang… - Proceedings of the …, 2020 - openaccess.thecvf.com
Automatic captioning of images is a task that combines the challenges of image analysis
and text generation. One important aspect of captioning is the notion of attention: how to …

Integrating tree path in transformer for code representation

H Peng, G Li, W Wang, Y Zhao… - Advances in Neural …, 2021 - proceedings.neurips.cc
Learning distributed representation of source code requires modelling its syntax and
semantics. Recent state-of-the-art models leverage highly structured source code …

Transformer grammars: Augmenting transformer language models with syntactic inductive biases at scale

L Sartran, S Barrett, A Kuncoro, M Stanojević… - Transactions of the …, 2022 - direct.mit.edu
Abstract We introduce Transformer Grammars (TGs), a novel class of Transformer language
models that combine (i) the expressive power, scalability, and strong performance of …

HiCLIP: Contrastive language-image pretraining with hierarchy-aware attention

S Geng, J Yuan, Y Tian, Y Chen, Y Zhang - arXiv preprint arXiv …, 2023 - arxiv.org
The success of large-scale contrastive vision-language pretraining (CLIP) has benefited
both visual recognition and multimodal content understanding. The concise design brings …