• Corpus ID: 226964679

Analyzing Sustainability Reports Using Natural Language Processing

@article{Luccioni2020AnalyzingSR,
  title={Analyzing Sustainability Reports Using Natural Language Processing},
  author={Alexandra Sasha Luccioni and Emily Baylor and Nicolas Anton Duch{\^e}ne},
  journal={ArXiv},
  year={2020},
  volume={abs/2011.08073}
}
Climate change is a far-reaching, global phenomenon that will impact many aspects of our society, including the global stock market \cite{dietz2016climate}. In recent years, companies have increasingly been aiming to both mitigate their environmental impact and adapt to the changing climate context. This is reported via increasingly exhaustive reports, which cover many types of climate risks and exposures under the umbrella of Environmental, Social, and Governance (ESG). However, given this… 

Tables from this paper

DEEPAGÉ: Answering Questions in Portuguese about the Brazilian Environment
TLDR
This work introduces multiple QA systems that combine in novel ways the BM25 algorithm, a sparse retrieval technique, with PTT5, a pre-trained state-of-the-art language model, focusing on the Portuguese language, thus offering resources not found elsewhere in the literature.
Tracking environmental policy changes in the Brazilian Federal Official Gazette
TLDR
A strategy to compose automated techniques and domain expert knowledge to process all the data from the Brazilian Federal Official Gazette is presented and the Government Actions Tracker, a highly curated dataset, in Portuguese, annotated by domain experts, on federal government acts about the Brazilian environmental policies is provided.
ClimateBert: A Pretrained Language Model for Climate-Related Text
TLDR
This work proposes CLIMATEBERT, a transformer-based language model that is further pretrained on over 1.6 million paragraphs of climate-related texts, crawled from various sources such as common news, research articles, and climate reporting of companies and finds that it leads to a 46% improvement on a masked language model objective.
Automated Identification of Climate Risk Disclosures in Annual Corporate Reports
TLDR
A dataset of over 120 manually-annotated annual reports by European firms is created to automatically identify disclosures of five different types of climate-related risks, finding that risk disclosure is increasing.
Tackling Climate Change with Machine Learning
TLDR
From smart grids to disaster management, high impact problems where existing gaps can be filled by ML are identified, in collaboration with other fields, to join the global effort against climate change.

References

SHOWING 1-10 OF 16 REFERENCES
Using Natural Language Processing to Analyze Financial Climate Disclosures
According to U.S. financial legislation, companies traded on the stock market are obliged to regularly disclose risks and uncertainties that are likely to affect their operations or financial
Tackling Climate Change with Machine Learning
TLDR
From smart grids to disaster management, high impact problems where existing gaps can be filled by ML are identified, in collaboration with other fields, to join the global effort against climate change.
Quantifying the Carbon Emissions of Machine Learning
TLDR
This work presents their Machine Learning Emissions Calculator, a tool for the community to better understand the environmental impact of training ML models and concrete actions that individual practitioners and organizations can take to mitigate their carbon emissions.
FinBERT: Financial Sentiment Analysis with Pre-trained Language Models
TLDR
FinBERT, a language model based on BERT, is introduced to tackle NLP tasks in the financial domain and it is found that even with a smaller training set and fine-tuning only a part of the model, FinBERT outperforms state-of-the-art machine learning methods.
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
TLDR
This article introduces BioBERT (Bidirectional Encoder Representations from Transformers for Biomedical Text Mining), which is a domain-specific language representation model pre-trained on large-scale biomedical corpora that largely outperforms BERT and previous state-of-the-art models in a variety of biomedical text mining tasks when pre- trained on biomedical Corpora.
‘Climate value at risk’ of global financial assets
Modelling shows that twenty-first-century climate change could significantly affect the market value of global financial assets, and suggests that limiting warming to no more than 2 °C would make
Supervised Learning of Universal Sentence Representations from Natural Language Inference Data
TLDR
It is shown how universal sentence representations trained using the supervised data of the Stanford Natural Language Inference datasets can consistently outperform unsupervised methods like SkipThought vectors on a wide range of transfer tasks.
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
TLDR
This work presents two parameter-reduction techniques to lower memory consumption and increase the training speed of BERT, and uses a self-supervised loss that focuses on modeling inter-sentence coherence.
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
TLDR
A new language representation model, BERT, designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.
...
...