Low-Rank Hidden State Embeddings for Viterbi Sequence Labeling
arXiv: Computation and Language, Volume abs/1708.00553, 2017.
EI
Abstract:
In textual information extraction and other sequence labeling tasks it is now common to use recurrent neural networks (such as LSTM) to form rich embedded representations of long-term input co-occurrence patterns. Representation of output co-occurrence patterns is typically limited to a hand-designed graphical model, such as a linear-chai...More
Code:
Data:
Get fulltext within 24h
Other Links
Upload PDF
1.Your uploaded documents will be check within 24h, and coins will be credited to your account.
2.As the current system does not support cash withdrawal, you can add staff WeChat (AMxiaomai) to receive it as a red packet.
3.10 coins will be exchanged for 1 yuan.
?
¥
Upload a single paper
for 5 coins
Wechat's Red Packet
?
¥
Upload 50 articles
for 280 coins
Wechat's Red Packet
?
¥
Upload 200 articles
for 1200 coins
Wechat's Red Packet
?
¥
Upload 500 articles
for 3000 coins
Wechat's Red Packet
?
¥
Upload 1000 articles
for 7000 coins
Wechat's Red Packet
Tags
Comments