@@ -121,8 +121,8 @@ def detect_phrases(self, tokenized_sentence: list[Token]) -> list[list[int]]:
121121
122122        text_pos  =  [(token .text , token .pos_tag ) for  token  in  self .tokenized_sentence ]
123123        parsed  =  self .mip_parser .parse (text_pos )
124-         for  subtree  in  parsed .subtrees (filter = lambda  t : t .label () in  ["EMIP" , "MIP" ]):
125-             indices  =  self ._get_subtree_indices (parsed , subtree )
124+         for  subtree  in  parsed .subtrees (filter = lambda  t : t .label () in  ["EMIP" , "MIP" ]):   #  type: ignore 
125+             indices  =  self ._get_subtree_indices (parsed , subtree )   #  type: ignore 
126126            # If the conjunction is not "or", skip 
127127            if  self ._cc_is_not_or (text_pos , indices ):
128128                continue 
@@ -172,8 +172,8 @@ def detect_sentences_splits(self, tokenized_sentence: list[Token]) -> list[int]:
172172
173173            text_pos .append ((t .feat_text , pos ))
174174        parsed  =  self .compound_parser .parse (text_pos )
175-         for  subtree  in  parsed .subtrees (filter = lambda  t : t .label () ==  "CS" ):
176-             indices  =  self ._get_subtree_indices (parsed , subtree )
175+         for  subtree  in  parsed .subtrees (filter = lambda  t : t .label () ==  "CS" ):   #  type: ignore 
176+             indices  =  self ._get_subtree_indices (parsed , subtree )   #  type: ignore 
177177            # If the conjunction is not "or", skip 
178178            if  self ._cc_is_not_or (text_pos , indices ):
179179                continue 
0 commit comments