CS代考 algorithm CSC485/2501 A1 Tutorial – cscodehelp代写
CSC485/2501 A1 Tutorial
TA: Zhewei Sun
▪ Due on Oct. 8th, at 11:59 pm.
▪ Asks you to implement a set of neural dependency parsers.
▪ Part 1: Transition-based dependency parser § See tutorial #1.
▪ Part 2: Graph-based dependency parser
▪ Part 1: Transition-based dependency parser § See tutorial #1.
▪ Part 2: Graph-based dependency parser § We’ll focus on this part today!
Errata for Tutorial #1
▪ The last slide presented in tutorial #1 mixed up the terms ‘projective’ and ‘non-projective’.
▪ Corrected: Transition-based parsers can only handle projective dependency parse trees and we’ll take a look at how to deal with non-projective cases today.
▪ Edge-factored Parsing Example
▪ Biaffine-attention Neural Dependency Parser • Arc scorer
• Label scorer
▪ Tips for Batching
Projective Dependency Trees
▪ Projectivity: If a dependency arc exist from i → j, then there exist a path from i → k for every min(i,j) < k < max(i,j). ▪ A dependency tree is projective if all of its arcs are projective. ▪ This corresponds to a planar graph: Non-projective Dependency Trees ▪ A non-projective dependency tree has one or more non-projective dependency arc. ▪ This corresponds to a graph with a cross-over arc: ▪ Transition-based parsers cannot handle this. Graph-based Dependency Parsing ▪ Consider all possible arcs between each pair of words: Edge-factored Parsing ▪ An edge-factored parser involves the following steps: 1. Create a bidirectional, connected graph that corresponds to the sentence. 2. Score each arc in the graph. 3. Finding the maximum spanning tree in the graph and treat the constituent arcs as your dependency arcs. 4. Tag each arc with an appropriate dependency label. Edge-factored Parsing ▪ Step 1: Create a bidirectional, connected graph that corresponds to the sentence. ▪ Input: Output: ▪ Add ROOT if necessary. Edge-factored Parsing ▪ Step 2: Score each edge in the graph. ▪ Input: Output: ▪ We will do this using a neural network (see the Arc Scorer). Edge-factored Parsing ▪ Step 3: Find the maximum spanning tree in the graph. ▪ Input: Output: Edge-factored Parsing ▪ Find the MST using Chu- algorithm. ▪ Pseudo-code from the textbook and example in the lecture. ▪ The MST can then be viewed as an unlabeled dependency parse tree: → Edge-factored Parsing ▪ Step 4: Tag each arc with an appropriate dependency label. ▪ Input: Output: ▪ We will also train a neural network to do this (see the Label Scorer). Biaffine-attention Neural Dependency Parser ▪ We need to train two neural networks to complete our parser: 1. Arc Scorer: Scores the likelihood for each arc in the graph. § The Scores are then passed to an MST algorithm. § Why can’t we just take the argmax here? 2. Label Scorer: Given an arc in the graph, output a score distribution over all possible tags. § In this case, we can simply take the argmax. ▪ For the original architecture, see Dozat et al. (2017): • https://aclanthology.org/K17-3002.pdf Contextualized Word Embedding ▪ Quick aside: The graph structure does not encode word order information. ▪ We can retain this information by using contextualized word embeddings, which is aware of its usage context: Word2Vec: BERT: Arc Scorer ▪ S – Contextual word embeddings for the sentence. ▪ DA – Learned representation for all words in the sentence when considered as dependents in arcs. ▪ HA – Learned representation for all words in the sentence when considered as heads in arcs. ▪ aij – Numeric score corresponding to the arc j → i Arc Scorer ▪ WA and bA are learnable parameters in the model. ▪ Use torch.nn.Parameter to create these weights/biases. – E.g. b_A = torch.nn.Parameter(torch.ones((10)), requires_grad=True) ▪ This tells PyTorch to register the tensor with your neural network module, in which you can find the parameters using self.parameters. Arc Scorer ▪ For this assignment, create the multi-layer perceptron (a.k.a. fully connected layers) using torch.nn.Sequential: – E.g. nn.Sequential(nn.Linear(50, 100), F.ReLU(), F.Dropout(0.5, training=True)) ▪ This creates a function layer that is equivalent to passing an input into the three layers sequentially. Arc Scorer ▪ The obtained score is then passed into a softmax layer: Arc Scorer ▪ Intuitions for the architecture: – The MLP layers create a pair of distinct representations for each word in the sentence, one for when the word is viewed as a dependent in an arc, and another for when it is viewed as a head. – The biaffine transformation between these representations then checks the appropriateness of having a directed edge between two words. – The bias term acts as a prior checking how likely a word is to be used as a head of an arc. Label Scorer ▪ S – Contextual word embeddings for the sentence. ▪ DL – Learned representation for all words in the sentence when considered as dependents in arcs. ▪ HL – Learned representation for all words in the sentence when considered as heads in arcs. ▪ lij – Vector of numeric scores where the k’th value correspond to the score of having the k’th dependency relation labeled for the arc j → i Label Scorer ▪ Since lij is a vector, WL and bL both have one extra dimension. – Back in the arc scorer case, W is of shape [X, X], but now it is of shape [X, Y, X], where X and Y are both integers. – The bias terms are now also vectors instead of a number. Label Scorer ▪ The score vector is then also passed into a softmax layer: ▪ Taking an argmax over the softmax scores gives the predicted label. Label Scorer ▪ Intuitions for the architecture: – The MLP layers and the biaffine transformation are very similar, except now that the biaffine transformation checks the appropriateness of having each individual tag on a directed edge between two words. – The two subsequent terms in the sum checks the appropriateness of having a certain word as the head/dependent of a given dependency relation. – The bias term sets a prior over all possible dependency relations. Batching ▪ We apply batched computation to take advantage of the powerful parallelism provided by modern hardware (i.e. GPUs). ▪ The tensors given to you in the assignment are already in batch format. ▪ You’ll need to maintain the given batch format in all neural network computations. Batching ▪ Directly encoding this into an array gives us: ▪ Cannot perform parallel computation because the sequence lengths are not matched. Batching ▪ Trick: Pad the shorter sequences and keep track of sequence lengths. ▪ This results in an extra batch dimension B in your tensor. In the example above, B=3. Working with Batched Tensors ▪ In order to perform efficient computation, you don’t want to unpack the batch dimension. ▪ Let A, B be two batched tensors of shape (b, X, Y) and (b, Y, Z) respectively. Now, to compute a matrix multiplication: Working with Batched Tensors ▪ torch.einsum(...) allows efficient batched operations. ▪ Separated by commas, each term to the left of the arrow defines the dimensions of the input tensors. ▪ The term following the arrow specifies the output, where every missing dimension will be summed over. Working with Batched Tensors ▪ torch.einsum(...) allows efficient batched operations. ▪ The matrix operation defined above above is thus equivalent to computing: result[b,i,k] = Σj A[b,i,j] * B[b,j,k] ∀ b, i, k ▪ More examples in the PyTorch documentation: § https://pytorch.org/docs/stable/generated/torch.einsum.html The End ▪ Good luck on the assignment! ▪ Please post any questions you have on Piazza.