This document summarizes a neural constituency parser that uses a transformer encoder and BiLSTM decoder. It achieves state-of-the-art 93.55 F1 score on the Penn Treebank dataset. The parser takes as input word embeddings, POS tags, and position embeddings. Ablation studies show that position embeddings contribute most to performance. The document also analyzes the model's use of attention heads and explores decomposing the input and attention mechanisms to improve parsing accuracy.