Program

The 4th Workshop on "Evaluation & Comparison of NLP Systems" 1st November 2023, co-located at IJCNLP-AACL 2023

Program Schedule

The workshop will take place virtually. Times on this page are in Central Indonesian time zone (UTC+8).

  • 9:00: Welcome and invited talk
  • 9:45: Session 1 (3 papers):
    • EduQuick: A Dataset Toward Evaluating Summarization of Informal Educational Content for Social Media
    • Assessing Distractors in Multiple-Choice Tests
    • Can a Prediction's Rank Offer a More Accurate Quantification of Bias? A Case Study Measuring Sexism in Debiased Language Models
  • 10:30: 5 minutes break
  • 10:35: Session 2 (3 papers):
    • Large Language Models As Annotators: A Preliminary Evaluation For Annotating Low-Resource Language Content
    • Summary Cycles: Exploring the Impact of Prompt Engineering on Large Language Models’ Interaction with Interaction Log Information
    • Transformers Go for the LOLs: Generating (Humourous) Titles from Scientific Abstracts End-to-End
  • 11:20: 5 minutes break
  • 11:25: Session 3 (3 papers):
    • Zero-shot Probing of Pretrained Language Models for Geography Knowledge
    • WRF: Weighted Rouge-F1 Metric for Entity Recognition
    • Delving into Evaluation Metrics for Generation: A Thorough Assessment of How Metrics Generalize to Rephrasing Across Languages
  • 12:15: 5 minutes break
  • 12:20: Session 4 (shared task)
    • 12:20-12:25: Shard Task Introduction
    • 12:30-12:50: Paper presentations
    • 12:50-13:00: Results and announcements of winners
  • 13:00: Wrap up and closing