-
Notifications
You must be signed in to change notification settings - Fork 7
Description
Hi @RunxinXu,
It seems that amrparse.py takes the transferred RAMS dataset as the input, which is similar to the WikiEvent dataset. However, there's no script for transferring the RAMS dataset. I tried to use the original RAMS dataset as the input of amrparse.py. but it seems that some sentences are too long, which causes an IndexError. Here's the detail of the error
Traceback (most recent call last):
File "/home/kxie/dh39/kyxie/transition-amr-parser-master/parser/amrparse.py", line 42, in
parse_rams(parser, split)
File "/home/kxie/dh39/kyxie/transition-amr-parser-master/parser/amrparse.py", line 17, in parse_rams
amr_list = parser.parse_sentences(all_sentences)
File "/home/kxie/dh39/kyxie/transition-amr-parser-master/transition_amr_parser/stack_transformer_amr_parser.py", line 422, in parse_sentences
roberta_batch_size)
File "/home/kxie/dh39/kyxie/transition-amr-parser-master/transition_amr_parser/stack_transformer_amr_parser.py", line 297, in convert_sentences_to_data
self.get_bert_features_batched(sentences, roberta_batch_size)
File "/home/kxie/dh39/kyxie/transition-amr-parser-master/transition_amr_parser/stack_transformer_amr_parser.py", line 270, in get_bert_features_batched
batch_data = self.embeddings.extract_batch(batch)
File "/home/kxie/dh39/kyxie/transition-amr-parser-master/transition_amr_parser/stack_transformer/pretrained_embeddings.py", line 310, in extract_batch
word_features = get_average_embeddings(roberta_features.unsqueeze(0), word2piece)
File "/home/kxie/dh39/kyxie/transition-amr-parser-master/transition_amr_parser/stack_transformer/pretrained_embeddings.py", line 21, in get_average_embeddings
column = final_layer[0:1, wordpiece_idx, :]
IndexError: index 510 is out of bounds for dimension 0 with size 510
In terms of getting the AMR result of the original RAMS dataset, do you have any suggestions?