Capturing alternative data in practice: How can neural networks understand news sentiment?

Dr Axel Groß-Klußmann in Quoniam’s team Research Forecasts set out to find the best approach to analysing daily news sentiment scores for over 1,000 economic topics. The neural network models are competitive when it comes to explaining stock market movements, currency changes, bond yields and predicting economic growth. His work has recently been published in Digital Finance.

Axel, you recently published your research “Learning deep news sentiment representations for macro-finance” in Digital Finance. What motivated you to explore this topic?

Whenever we have tried to bring together alternative sentiment data and financial time series, the biggest obstacle has always been the selection or weighting of the underlying narratives or topics. Some form of dimension reduction is required to link news collected on potentially thousands of topics to individual financial instruments. However, most of the standard statistical approaches left something to be desired. Often, too many data were discarded while other approaches produced weightings for topics irrespective of our use case context.

Given that neural networks are strong at these kinds of tasks, I was simply pursuing an idea I had about how to use neural networks to give us what we wanted. Recent software advances have made it very easy to customise and train neural networks, so it was just a question of seeing how the strengths of neural networks would carry over to our finance use case with considerably fewer data points than are typically used for training these models.

How did you approach these questions in detail?

I started the analysis using the standard dimension reduction methods like principle component analysis (PCA). However, it quickly became clear that due to the complexity of news sentiment data we needed some form of supervision when reducing the dimensionality. So we added partial least squares (PLS), regression methods such as ridge and lasso, or supervised PCAs approaches. Still, I felt that we could capture complex features of news data better with more expressive statistical approaches, which led directly to the autoencoder branch of neural networks.

Several software solutions in this space are basically at your fingertips these days, which means you can implement a neural network in no time. Tech-wise, I employed PyTorch, a popular open-source deep learning package. In this framework I coded up a custom architecture for autoencoders with supervisions (supervised autoencoders) that integrates both dimensionality reduction and supervised learning. In simple words, this particular architecture allows the model to learn compact sentiment data representations that are already well adapted to our financial use case by way of the supervisions.

What were your findings?

On the positive side, training and coding of the neural networks was much easier than originally thought. Above all, the analysis highlighted the added value of customisation in statistical signal construction. The implementation of custom loss functions in particular resulted in data representations that we liked much better than those of off-the-shelf approaches. Across the board, the results in different use cases robustly favoured the neural networks. Interestingly, the complexity of the neural networks in terms of the depth was of second order. Also, while it was hard to get something significantly better than standard approaches after accounting for multiple selection issues etc., the neural networks produced data representations that were more relevant to us practitioners.

How do you apply these insights at Quoniam in the investment process?

While we are looking at including neural network derived data representations into investment strategies, the proposed approach will be at first most helpful in the research process itself. One example is the review of datasets containing text-derived numerical sentiment scores for many topics. The framework outlined in the paper gives us a very competitive benchmark approach that can directly produce actionable signals.

YOU MIGHT ALSO BE INTERESTED IN
Article
September 2024
How low volatility boosts compounded returns – case study emerging markets

The low volatility anomaly explains the advantages of low volatility investing. An additional benefit: The lower the volatility, the greater the compounding of portfolio returns. In their new white paper, Carsten Rother and Dr Xavier Gerard, CFA, find that using a low volatility approach in emerging markets can significantly reduce the risk of the market portfolio, making the compounding effect exceptionally powerful.

Interview
January 2024
Bonds with benefits: Combining sustainability and return potential in corporate bond portfolios

Can investors in corporate bonds do good and earn an attractive return at the same time? That was what Quoniam experts Dr Desislava Vladimirova and Dr Jieyan Fang-Klinger set out to determine in their recent research paper, Bonds with Benefits: Impact Investing in Corporate Debt, recently published in the Financial Analyst Journal. We spoke to them about their research, what they found and what it means for investors.

Video
October 2023
Asset Pricing: Missing Data and the Quality of Signals – talk with Michael Weber and Volker Flögel

What are the most common pitfalls that practitioners face when dealing with missing data? And is it possible to extend the history of data by backfilling datasets, for example with ESG data where there is not so much reliable historical data? Find out in this insightful discussion with Michael Weber, Associate Professor at the University of Chicago Booth School of Business, and Volker Flögel, CIO at Quoniam. Together they discuss how to deal with missing data in asset pricing and investment – also in light of the current market environment – and give an outlook on upcoming research in this area.