Character-level Convolutional Neural
Network for Sentence Paraphrase Detection
Vladislav Maraev
NLX-Group, Faculty of Sciences, University of Lisbon
Paraphrase detection for Russian workshop
AINL FRUCT 2016
Objective
What
Task 2 — Binary classification (paraphrase/non-paraphrase).
How
Apply convolutional neural network (CNN) architecture:
Standard Non-standard
Word embeddings ✓ ✓
Character embeddings ✓
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 2 / 16
Related work
Convolutional neural networks in NLP
• Detecting semantically equivalent questions with CNN and
word embeddings (Bogdanova et al., 2015)
• Convolutional Neural Networks for Sentence Classification
(Zhang and Wallace, 2015)
• Attention-based CNN for modeling sentence pairs (Yin et al.,
2016)
• Character embeddings for text classification(Zhang et al.,
2015)
• Word+character embeddings for sentiment analysis (dos
Santos and Gatti, 2014)
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 3 / 16
How CNN works?
TR CONV POOL cosine
similarity
Steps:
1. Token representation (Embedding)
2. Convolution
3. Pooling
4. Pair similarity estimation
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 4 / 16
Convolutional Neural Network
1. Token representation
Input
s = {t1, t2, . . . , tN}
Token representation
rt
= W0
vt
, (1)
where
• W0 ∈ Rd×V is an embedding matrix
• vt ∈ RV is a one-hot encoded vector of size V
Output
sTR = {rt1 , rt2 , . . . , rtN } , where rtn ∈ Rd
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 5 / 16
Convolutional Neural Network
2. Convolution
Convolution
1. Concatenations zn of k-grams
2. Multiply by W1, add bias b1, and
apply tanh function:
rzn
= tanh
(
W1
zn + b1
)
where:
zn ∈ Rdk
W1
∈ Rclu×dk
rzn
∈ Rclu
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 6 / 16
Convolutional Neural Network
3. Pooling
Sum (or Max) over all the rzn element-wise and apply tanh
function:
rs
= tanh
(
∑
n
rzn
)
which will give us sentence representation rs ∈ Rclu
* This means that sentence representation doesn’t depend on
sentence length.
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 7 / 16
Convolutional Neural Network
4. Compute similarity
TR CONV POOL cosine
similarity
Estimate similarity between the pair of sentence
representations using cosine measure:
similarity =
rs1 · rs2
∥rs1 ∥∥rs2 ∥
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 8 / 16
Training the network
We train W0, W1 and b1.
Steps
1. Compute mean-squared error (w.r.t. cosine similarity)
2. Use the backpropagation algorithm (SGD/RMSProp) to
compute gradients of the network
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 9 / 16
Several convolutional filters
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 10 / 16
Standard Run Hyperparameters
Word embeddings
Parameter Value Description
k {3, 5, 8, 12} Sizes of k-grams
clu 100 Size of each convolutional filter
d 300 Size of word representation
epochs 5 Number of training epochs
pooling MAX pooling layer function
optimiser RMSProp Keras’s optimiser
word embeddings Random (uniform)
Sentences were tokenised and lowercased using Keras.
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 11 / 16
Standard Run Hyperparameters
Character embeddings
Parameter Value Description
k {2, 3, 5, 7, 9, 11} Sizes of k-grams
clu 100 Size of each conv. filter
d 100 Size of word representation
epochs 20 Number of training epochs
pooling MAX pooling layer function
optimiser RMSProp Keras’s optimiser
char. embeddings Random (uniform)
Characters were lowercased, non-word characters were removed.
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 12 / 16
Non-Standard Run Hyperparameters
Parameter Value Description
k 3 Size of k-gram
clu 300 Size of convolutional filter
d 300 Size of word representation
epochs 5 Number of training epochs
pooling MAX pooling layer function
optimiser RMSProp Keras’s optimiser
word embeddings RusVect¯or¯es trained on Russian National Corpus
(Kutuzov and Andreev, 2015)
Input sentences were tokenised, lemmatised and PoS-tagged
with MyStem (Segalovich, 2003).
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 13 / 16
Main results
Accuracy F1
Standard
NLX (characters) 72.74 78.80
NLX (words) 66.19 76.44
Non-standard NLX (words) 69.94 76.80
BASELINE 49.66 54.03
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 14 / 16
Discussion
1. The result for standard run is competing with the best system
and can be further improved by tuning hyperparameters
automatically and also picking the epoch for testing
automatically, based on the validation results.
2. Surprisingly, results for the standard run outperformed
non-standard, however, non-standard used external resources
for lemmatisation and initial word embeddings. (Probably due
to a higher focus on the standard run).
Next? Attention-based CNN (Yin et al., 2016), combination of
character and word embeddings (dos Santos and Gatti, 2014).
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 15 / 16
References
Dasha Bogdanova, Cıcero dos Santos, Luciano Barbosa, and Bianca Zadrozny.
Detecting semantically equivalent questions in online user forums. CoNLL 2015,
page 123, 2015.
Cicero dos Santos and Maira Gatti. Deep convolutional neural networks for sentiment
analysis of short texts. In Proceedings of COLING 2014, the 25th International
Conference on Computational Linguistics: Technical Papers, pages 69–78, Dublin,
Ireland, August 2014. Dublin City University and Association for Computational
Linguistics.
Andrey Kutuzov and Igor Andreev. Texts in, meaning out: neural language models in
semantic similarity task for russian. In Proceedings of the Dialog Conference, 2015.
Ilya Segalovich. A fast morphological algorithm with unknown word guessing induced
by a dictionary for a web search engine. In MLMTA, pages 273–280. Citeseer, 2003.
Wenpeng Yin, Hinrich Schtze, Bing Xiang, and Bowen Zhou. Abcnn: Attention-based
convolutional neural network for modeling sentence pairs. Transactions of the
Association for Computational Linguistics, 4:259–272, 2016. ISSN 2307-387X.
Xiang Zhang, Junbo Zhao, and Yann LeCun. Character-level convolutional networks
for text classification. In Advances in Neural Information Processing Systems,
pages 649–657, 2015.
Ye Zhang and Byron Wallace. A sensitivity analysis of (and practitioners’ guide to)
convolutional neural networks for sentence classification. arXiv preprint
arXiv:1510.03820, 2015.
Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 16 / 16

More Related Content

PDF
AINL 2016: Galinsky, Alekseev, Nikolenko
PDF
AINL 2016: Alekseev, Nikolenko
PDF
AINL 2016: Eyecioglu
PDF
AINL 2016: Kravchenko
PDF
AINL 2016: Kuznetsova
PPTX
AINL 2016: Yagunova
PDF
Crash Course in Natural Language Processing (2016)
PDF
AINL 2016: Bastrakova, Ledesma, Millan, Zighed
AINL 2016: Galinsky, Alekseev, Nikolenko
AINL 2016: Alekseev, Nikolenko
AINL 2016: Eyecioglu
AINL 2016: Kravchenko
AINL 2016: Kuznetsova
AINL 2016: Yagunova
Crash Course in Natural Language Processing (2016)
AINL 2016: Bastrakova, Ledesma, Millan, Zighed

What's hot (20)

PDF
Crash-course in Natural Language Processing
PPTX
Language models
PDF
Can functional programming be liberated from static typing?
PDF
NLP Project Full Cycle
PDF
Aspects of NLP Practice
PPTX
Detecting and Describing Historical Periods in a Large Corpora
PDF
Natural Language Processing in Practice
PDF
The State of #NLProc
PDF
Practical NLP with Lisp
PDF
Language Models for Information Retrieval
PDF
Using Linked Disambiguated Distributional Networks for Word Sense Disambiguation
PDF
OUTDATED Text Mining 5/5: Information Extraction
PDF
Overview of text mining and NLP (+software)
PDF
Lodifier: Generating Linked Data from Unstructured Text
PDF
NEEL2015 challenge summary
PDF
IRJET- Automatic Language Identification using Hybrid Approach and Classifica...
PDF
DeepWalk: Online Learning of Representations
PDF
Frontiers of Natural Language Processing
PPTX
The Duet model
PDF
OUTDATED Text Mining 3/5: String Processing
Crash-course in Natural Language Processing
Language models
Can functional programming be liberated from static typing?
NLP Project Full Cycle
Aspects of NLP Practice
Detecting and Describing Historical Periods in a Large Corpora
Natural Language Processing in Practice
The State of #NLProc
Practical NLP with Lisp
Language Models for Information Retrieval
Using Linked Disambiguated Distributional Networks for Word Sense Disambiguation
OUTDATED Text Mining 5/5: Information Extraction
Overview of text mining and NLP (+software)
Lodifier: Generating Linked Data from Unstructured Text
NEEL2015 challenge summary
IRJET- Automatic Language Identification using Hybrid Approach and Classifica...
DeepWalk: Online Learning of Representations
Frontiers of Natural Language Processing
The Duet model
OUTDATED Text Mining 3/5: String Processing
Ad

Viewers also liked (20)

PDF
AINL 2016: Strijov
PDF
AINL 2016: Skornyakov
PDF
AINL 2016: Goncharov
PPT
AINL 2016: Bodrunova, Blekanov, Maksimov
PPTX
AINL 2016: Ustalov
PDF
AINL 2016: Proncheva
PDF
AINL 2016: Romanova, Nefedov
PPTX
AINL 2016: Bugaychenko
PDF
AINL 2016: Castro, Lopez, Cavalcante, Couto
PDF
AINL 2016: Fenogenova, Karpov, Kazorin
PDF
AINL 2016: Panicheva, Ledovaya
PPTX
AINL 2016: Rykov, Nagornyy, Koltsova, Natta, Kremenets, Manovich, Cerrone, Cr...
PDF
AINL 2016: Nikolenko
PDF
AINL 2016: Boldyreva
PDF
AINL 2016: Kozerenko
PDF
AINL 2016: Muravyov
PPTX
AINL 2016: Moskvichev
PDF
AINL 2016: Khudobakhshov
PDF
AINL 2016: Malykh
PDF
AINL 2016: Filchenkov
AINL 2016: Strijov
AINL 2016: Skornyakov
AINL 2016: Goncharov
AINL 2016: Bodrunova, Blekanov, Maksimov
AINL 2016: Ustalov
AINL 2016: Proncheva
AINL 2016: Romanova, Nefedov
AINL 2016: Bugaychenko
AINL 2016: Castro, Lopez, Cavalcante, Couto
AINL 2016: Fenogenova, Karpov, Kazorin
AINL 2016: Panicheva, Ledovaya
AINL 2016: Rykov, Nagornyy, Koltsova, Natta, Kremenets, Manovich, Cerrone, Cr...
AINL 2016: Nikolenko
AINL 2016: Boldyreva
AINL 2016: Kozerenko
AINL 2016: Muravyov
AINL 2016: Moskvichev
AINL 2016: Khudobakhshov
AINL 2016: Malykh
AINL 2016: Filchenkov
Ad

Similar to AINL 2016: Maraev (20)

PDF
IRJET - Analysis of Paraphrase Detection using NLP Techniques
PDF
Grammarly Meetup: Paraphrase Detection in NLP (PART 2) - Andriy Gryshchuk
PPTX
Talk from NVidia Developer Connect
PPTX
Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...
PPTX
NLTK - Natural Language Processing in Python
PDF
UCU NLP Summer Workshops 2017 - Part 2
PDF
Deep learning for nlp
PDF
Dli meetup-may 27-milano_nlp_de_nobili
PDF
Semeval Deep Learning In Semantic Similarity
PDF
Generating Natural-Language Text with Neural Networks
PDF
Seq2seq Model to Tokenize the Chinese Language
PDF
Seq2seq Model to Tokenize the Chinese Language
PDF
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-b...
PPTX
Dcnn for text
PDF
Neural Machine Translation (D2L10 Insight@DCU Machine Learning Workshop 2017)
PDF
Let Android dream electric sheep: Making emotion model for chat-bot with Pyth...
PDF
Neural Machine Translation (D3L4 Deep Learning for Speech and Language UPC 2017)
PDF
Bay Area NLP Reading Group - 7.12.16
PDF
IRJET- Visual Information Narrator using Neural Network
IRJET - Analysis of Paraphrase Detection using NLP Techniques
Grammarly Meetup: Paraphrase Detection in NLP (PART 2) - Andriy Gryshchuk
Talk from NVidia Developer Connect
Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...
NLTK - Natural Language Processing in Python
UCU NLP Summer Workshops 2017 - Part 2
Deep learning for nlp
Dli meetup-may 27-milano_nlp_de_nobili
Semeval Deep Learning In Semantic Similarity
Generating Natural-Language Text with Neural Networks
Seq2seq Model to Tokenize the Chinese Language
Seq2seq Model to Tokenize the Chinese Language
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-b...
Dcnn for text
Neural Machine Translation (D2L10 Insight@DCU Machine Learning Workshop 2017)
Let Android dream electric sheep: Making emotion model for chat-bot with Pyth...
Neural Machine Translation (D3L4 Deep Learning for Speech and Language UPC 2017)
Bay Area NLP Reading Group - 7.12.16
IRJET- Visual Information Narrator using Neural Network

More from Lidia Pivovarova (8)

PDF
Classification and clustering in media monitoring: from knowledge engineering...
PDF
Convolutional neural networks for text classification
PDF
Grouping business news stories based on salience of named entities
PDF
Интеллектуальный анализ текста
PDF
AINL 2016: Shavrina, Selegey
PPTX
AINL 2016:
PDF
AINL 2016: Grigorieva
PDF
AINL 2016: Just AI
Classification and clustering in media monitoring: from knowledge engineering...
Convolutional neural networks for text classification
Grouping business news stories based on salience of named entities
Интеллектуальный анализ текста
AINL 2016: Shavrina, Selegey
AINL 2016:
AINL 2016: Grigorieva
AINL 2016: Just AI

Recently uploaded (20)

PPTX
Cutaneous tuberculosis Dermatology
PPTX
Chapter 1 Introductory course Biology Camp
PDF
final prehhhejjehehhehehehebesentation.pdf
PDF
Chemistry and Changes 8th Grade Science .pdf
PPTX
Cells and Organs of the Immune System (Unit-2) - Majesh Sir.pptx
PDF
No dilute core produced in simulations of giant impacts on to Jupiter
PPTX
Posology_43998_PHCEUTICS-T_13-12-2023_43998_PHCEUTICS-T_17-07-2025.pptx
PDF
ECG Practice from Passmedicine for MRCP Part 2 2024.pdf
PPTX
The Female Reproductive System - Grade 10 ppt
PDF
Unit Four Lesson in Carbohydrates chemistry
PPT
ecg for noob ecg interpretation ecg recall
PPT
ZooLec Chapter 13 (Digestive System).ppt
PDF
Integrative Oncology: Merging Conventional and Alternative Approaches (www.k...
PDF
Thyroid Hormone by Iqra Nasir detail.pdf
PDF
TOPIC-1-Introduction-to-Bioinformatics_for dummies
PPT
Chapter 6 Introductory course Biology Camp
PPTX
Targeted drug delivery system 1_44299_BP704T_03-12-2024.pptx
PDF
Traditional Healing Practices: A Model for Integrative Care in Diabetes Mana...
PPTX
Models of Eucharyotic Chromosome Dr. Thirunahari Ugandhar.pptx
PDF
The Physiology Of The Red Blood Cells pdf
Cutaneous tuberculosis Dermatology
Chapter 1 Introductory course Biology Camp
final prehhhejjehehhehehehebesentation.pdf
Chemistry and Changes 8th Grade Science .pdf
Cells and Organs of the Immune System (Unit-2) - Majesh Sir.pptx
No dilute core produced in simulations of giant impacts on to Jupiter
Posology_43998_PHCEUTICS-T_13-12-2023_43998_PHCEUTICS-T_17-07-2025.pptx
ECG Practice from Passmedicine for MRCP Part 2 2024.pdf
The Female Reproductive System - Grade 10 ppt
Unit Four Lesson in Carbohydrates chemistry
ecg for noob ecg interpretation ecg recall
ZooLec Chapter 13 (Digestive System).ppt
Integrative Oncology: Merging Conventional and Alternative Approaches (www.k...
Thyroid Hormone by Iqra Nasir detail.pdf
TOPIC-1-Introduction-to-Bioinformatics_for dummies
Chapter 6 Introductory course Biology Camp
Targeted drug delivery system 1_44299_BP704T_03-12-2024.pptx
Traditional Healing Practices: A Model for Integrative Care in Diabetes Mana...
Models of Eucharyotic Chromosome Dr. Thirunahari Ugandhar.pptx
The Physiology Of The Red Blood Cells pdf

AINL 2016: Maraev

  • 1. Character-level Convolutional Neural Network for Sentence Paraphrase Detection Vladislav Maraev NLX-Group, Faculty of Sciences, University of Lisbon Paraphrase detection for Russian workshop AINL FRUCT 2016
  • 2. Objective What Task 2 — Binary classification (paraphrase/non-paraphrase). How Apply convolutional neural network (CNN) architecture: Standard Non-standard Word embeddings ✓ ✓ Character embeddings ✓ Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 2 / 16
  • 3. Related work Convolutional neural networks in NLP • Detecting semantically equivalent questions with CNN and word embeddings (Bogdanova et al., 2015) • Convolutional Neural Networks for Sentence Classification (Zhang and Wallace, 2015) • Attention-based CNN for modeling sentence pairs (Yin et al., 2016) • Character embeddings for text classification(Zhang et al., 2015) • Word+character embeddings for sentiment analysis (dos Santos and Gatti, 2014) Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 3 / 16
  • 4. How CNN works? TR CONV POOL cosine similarity Steps: 1. Token representation (Embedding) 2. Convolution 3. Pooling 4. Pair similarity estimation Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 4 / 16
  • 5. Convolutional Neural Network 1. Token representation Input s = {t1, t2, . . . , tN} Token representation rt = W0 vt , (1) where • W0 ∈ Rd×V is an embedding matrix • vt ∈ RV is a one-hot encoded vector of size V Output sTR = {rt1 , rt2 , . . . , rtN } , where rtn ∈ Rd Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 5 / 16
  • 6. Convolutional Neural Network 2. Convolution Convolution 1. Concatenations zn of k-grams 2. Multiply by W1, add bias b1, and apply tanh function: rzn = tanh ( W1 zn + b1 ) where: zn ∈ Rdk W1 ∈ Rclu×dk rzn ∈ Rclu Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 6 / 16
  • 7. Convolutional Neural Network 3. Pooling Sum (or Max) over all the rzn element-wise and apply tanh function: rs = tanh ( ∑ n rzn ) which will give us sentence representation rs ∈ Rclu * This means that sentence representation doesn’t depend on sentence length. Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 7 / 16
  • 8. Convolutional Neural Network 4. Compute similarity TR CONV POOL cosine similarity Estimate similarity between the pair of sentence representations using cosine measure: similarity = rs1 · rs2 ∥rs1 ∥∥rs2 ∥ Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 8 / 16
  • 9. Training the network We train W0, W1 and b1. Steps 1. Compute mean-squared error (w.r.t. cosine similarity) 2. Use the backpropagation algorithm (SGD/RMSProp) to compute gradients of the network Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 9 / 16
  • 10. Several convolutional filters Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 10 / 16
  • 11. Standard Run Hyperparameters Word embeddings Parameter Value Description k {3, 5, 8, 12} Sizes of k-grams clu 100 Size of each convolutional filter d 300 Size of word representation epochs 5 Number of training epochs pooling MAX pooling layer function optimiser RMSProp Keras’s optimiser word embeddings Random (uniform) Sentences were tokenised and lowercased using Keras. Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 11 / 16
  • 12. Standard Run Hyperparameters Character embeddings Parameter Value Description k {2, 3, 5, 7, 9, 11} Sizes of k-grams clu 100 Size of each conv. filter d 100 Size of word representation epochs 20 Number of training epochs pooling MAX pooling layer function optimiser RMSProp Keras’s optimiser char. embeddings Random (uniform) Characters were lowercased, non-word characters were removed. Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 12 / 16
  • 13. Non-Standard Run Hyperparameters Parameter Value Description k 3 Size of k-gram clu 300 Size of convolutional filter d 300 Size of word representation epochs 5 Number of training epochs pooling MAX pooling layer function optimiser RMSProp Keras’s optimiser word embeddings RusVect¯or¯es trained on Russian National Corpus (Kutuzov and Andreev, 2015) Input sentences were tokenised, lemmatised and PoS-tagged with MyStem (Segalovich, 2003). Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 13 / 16
  • 14. Main results Accuracy F1 Standard NLX (characters) 72.74 78.80 NLX (words) 66.19 76.44 Non-standard NLX (words) 69.94 76.80 BASELINE 49.66 54.03 Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 14 / 16
  • 15. Discussion 1. The result for standard run is competing with the best system and can be further improved by tuning hyperparameters automatically and also picking the epoch for testing automatically, based on the validation results. 2. Surprisingly, results for the standard run outperformed non-standard, however, non-standard used external resources for lemmatisation and initial word embeddings. (Probably due to a higher focus on the standard run). Next? Attention-based CNN (Yin et al., 2016), combination of character and word embeddings (dos Santos and Gatti, 2014). Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 15 / 16
  • 16. References Dasha Bogdanova, Cıcero dos Santos, Luciano Barbosa, and Bianca Zadrozny. Detecting semantically equivalent questions in online user forums. CoNLL 2015, page 123, 2015. Cicero dos Santos and Maira Gatti. Deep convolutional neural networks for sentiment analysis of short texts. In Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers, pages 69–78, Dublin, Ireland, August 2014. Dublin City University and Association for Computational Linguistics. Andrey Kutuzov and Igor Andreev. Texts in, meaning out: neural language models in semantic similarity task for russian. In Proceedings of the Dialog Conference, 2015. Ilya Segalovich. A fast morphological algorithm with unknown word guessing induced by a dictionary for a web search engine. In MLMTA, pages 273–280. Citeseer, 2003. Wenpeng Yin, Hinrich Schtze, Bing Xiang, and Bowen Zhou. Abcnn: Attention-based convolutional neural network for modeling sentence pairs. Transactions of the Association for Computational Linguistics, 4:259–272, 2016. ISSN 2307-387X. Xiang Zhang, Junbo Zhao, and Yann LeCun. Character-level convolutional networks for text classification. In Advances in Neural Information Processing Systems, pages 649–657, 2015. Ye Zhang and Byron Wallace. A sensitivity analysis of (and practitioners’ guide to) convolutional neural networks for sentence classification. arXiv preprint arXiv:1510.03820, 2015. Vladislav Maraev AINL FRUCT 2016 Paraphrase detection for Russian workshop (10.11.2016) 16 / 16