What is the most effective method to pool BERT embeddings for text classification? A case study in gender-responsive aid
Text embeddings are typically aggregated before use in classification. Does the method in which the embeddings are pooled affect the ease of classifying the text?