• Annotator rationales for labeling tasks in crowdsourcing 

      Kutlu, Mucahid; McDonnell, Tyler; Elsayed, Tamer; Lease, Matthew ( Elsevier , 2020 , Article)
      When collecting item ratings from human judges, it can be difficult to measure and enforce data quality due to task subjectivity and lack of transparency into how judges make each rating decision. To address this, we ...
    • ArabicWeb16: A new crawl for today's Arabic Web 

      Suwaileh, Reem; Kutlu, Mucahid; Fathima, Nihal; Elsayed, Tamer; Lease, Matthew ( Association for Computing Machinery, Inc , 2016 , Conference Paper)
      Web crawls provide valuable snapshots of the Web which enable a wide variety of research, be it distributional analysis to characterize Web properties or use of language, content analysis in social science, or Information ...
    • ArTest: The First Test Collection for Arabic Web Search with Relevance Rationales 

      Hasanain, Maram; Barkallah, Yassmine; Suwaileh, Reem; Kutlu, Mucahid; Elsayed, Tamer ( Association for Computing Machinery, Inc , 2020 , Conference Paper)
      The scarcity of Arabic test collections has long hindered information retrieval (IR) research over the Arabic Web. In this work, we present ArTest, the first large-scale test collection designed for the evaluation of ad-hoc ...
    • Crowd vs. Expert: What can relevance judgment rationales teach us about assessor disagreement? 

      Kutlu, M.; Kutlu, Mucahid; McDonnell, Tyler; Barkallah, Yassmine; Elsayed, Tamer; ... more authors ( ACM , 2018 , Conference Paper)
      © 2018 ACM. While crowdsourcing offers a low-cost, scalable way to collect relevance judgments, lack of transparency with remote crowd work has limited understanding about the quality of collected judgments. In prior work, ...
    • Efficient Test Collection Construction via Active Learning 

      Rahman, Md Mustafizur; Kutlu, Mucahid; Elsayed, Tamer; Lease, Matthew ( Association for Computing Machinery , 2020 , Conference Paper)
      To create a new IR test collection at low cost, it is valuable to carefully select which documents merit human relevance judgments. Shared task campaigns such as NIST TREC pool document rankings from many participating ...
    • The many benefits of annotator rationales for relevance judgments 

      McDonnell, Tyler; Kutlu, Mucahid; Elsayed, Tamer; Lease, Matthew ( International Joint Conferences on Artificial Intelligence , 2017 , Conference Paper)
      When collecting subjective human ratings of items, it can be difficult to measure and enforce data quality due to task subjectivity and lack of insight into how judges arrive at each rating decision. To address this, we ...
    • Mix and match: Collaborative expert-crowd judging for building test collections accurately and affordably 

      Kutlu, Mucahid; McDonnell, Tyler; Sheshadri, Aashish; Elsayed, Tamer; Lease, Matthew ( CEUR-WS , 2018 , Conference Paper)
      Crowdsourcing offers an affordable and scalable means to collect relevance judgments for information retrieval test collections. However, crowd assessors may showhigher variance in judgment quality than trusted assessors. ...
    • Overview of the CLEF-2021 CheckThat! Lab Task 1 on check-worthiness estimation in tweets and political debates 

      Shaar, Shaden; Hasanain, Maram; Hamdan, Bayan; Ali, Zien Sheikh; Haouari, Fatima; ... more authors ( CEUR-WS , 2021 , Conference Paper)
      We present an overview of Task 1 of the fourth edition of the CheckThat! Lab, part of the 2021 Conference and Labs of the Evaluation Forum (CLEF). The task asks to predict which posts in a Twitter stream are worth ...
    • Overview of the CLEF–2021 CheckThat! Lab on Detecting Check-Worthy Claims, Previously Fact-Checked Claims, and Fake News 

      Nakov, Preslav; Da San Martino, Giovanni; Elsayed, Tamer; Barrón-Cedeño, Alberto; Míguez, Rubén; ... more authors ( Springer Science and Business Media Deutschland GmbH , 2021 , Conference Paper)
      We describe the fourth edition of the CheckThat! Lab, part of the 2021 Conference and Labs of the Evaluation Forum (CLEF). The lab evaluates technology supporting tasks related to factuality, and covers Arabic, Bulgarian, ...
    • Overview of the CLEF–2023 CheckThat! Lab on Checkworthiness, Subjectivity, Political Bias, Factuality, and Authority of News Articles and Their Source 

      Barrón-Cedeño, Alberto; Alam, Firoj; Galassi, Andrea; Da San Martino, Giovanni; Nakov, Preslav; ... more authors ( Springer Science and Business Media Deutschland GmbH , 2023 , Conference Paper)
      We describe the sixth edition of the CheckThat! lab, part of the 2023 Conference and Labs of the Evaluation Forum (CLEF). The five previous editions of CheckThat! focused on the main tasks of the information verification ...
    • Par-eXpress: A tool for analysis of sequencing experiments with ambiguous assignment of fragments in parallel 

      Kutlu, Mucahid; Agrawal, Gagan; Blachly, James S. ( Institute of Electrical and Electronics Engineers Inc. , 2017 , Conference Paper)
      With new high-throughput and low-cost sequencing technologies, an increasing amount of genetic data is becoming available to researchers. While the analysis of this vast amount of data has great potential for future ...
    • Your Behavior Signals Your Reliability: Modeling Crowd Behavioral Traces to Ensure Quality Relevance Annotations 

      Goyal, Tanya; McDonnell, Tyler; Kutlu, Mucahid; Elsayed, Tamer; Lease, Matthew ( AAAI Press , 2018 , Conference Paper)
      While peer-agreement and gold checks are well-established methods for ensuring quality in crowdsourced data collection, we explore a relatively new direction for quality control: estimating work quality directly from ...