MAEC: A Multimodal Aligned Earnings Conference Call Dataset for Financial Risk Prediction

Jiazheng Li, Linyi Yang, Barry Smyth, Ruihai Dong

Research output: Chapter in Book/Report/Conference proceedingConference paperpeer-review

24 Citations (Scopus)

Abstract

In the area of natural language processing, various financial datasets have informed recent research and analysis including financial news, financial reports, social media, and audio data from earnings calls. We introduce a new, large-scale multi-modal, text-audio paired, earnings-call dataset named MAEC, based on S&P 1500 companies. We describe the main features of MAEC, how it was collected and assembled, paying particular attention to the text-audio alignment process used. We present the approach used in this work as providing a suitable framework for processing similar forms of data in the future. The resulting dataset is more than six times larger than those currently available to the research community and we discuss its potential in terms of current and future research challenges and opportunities. All resources of this work are available at https://github.com/Earnings-Call-Dataset/
Original languageEnglish
Title of host publicationProceedings of the 29th ACM International Conference on Information & Knowledge Management
Place of PublicationNew York, NY, USA
PublisherAssociation for Computing Machinery
Pages3063–3070
ISBN (Print)9781450368599
DOIs
Publication statusPublished - 2020

Publication series

NameCIKM '20
PublisherAssociation for Computing Machinery

Keywords

  • earnings conference calls
  • multimodal aligned datasets
  • financial risk prediction

Fingerprint

Dive into the research topics of 'MAEC: A Multimodal Aligned Earnings Conference Call Dataset for Financial Risk Prediction'. Together they form a unique fingerprint.

Cite this