Managing risk within the financial services, especially within the banking sector, has increased in complexity over the past several years.
First, new frameworks (such as FRTB) are being introduced that potentially require tremendous computing power and an ability to analyze years of historical data.
At the same, regulators are demanding more transparency and explainability from the banks they oversee.
Finally, the introduction of new technologies and business models means the need for sound risk governance is at an all time high.
However, the ability for the banking industry to effectively meet these demands has not been an easy undertaking.
Traditional banks relying on on-premises infrastructure can no longer effectively manage risk.
Banks must abandon the computational inefficiencies of legacy technologies and build an agile Modern Risk Management practice capable of rapidly responding to market and economic volatility through the use of data and advanced analytics.
Recent experience shows that as new threats emerge, historical data and aggregated risk models lose their predictive values quickly.
Risk analysts must augment traditional data with alternative datasets in order to explore new ways of identifying and quantifying the risks facing their business, both at scale and in real-time.
In this blog, we will demonstrate how to modernize traditional value-at-risk (VaR) calculation through the use of various components of the Databricks Unified Data Analytics Platform — Delta Lake, Apache SparkTM and MLflow — in order to enable a more agile and forward looking approach to risk management.
This first series of notebooks will cover the multiple data engineering and data science challenges that must be addressed to effectively modernize risk management practices: Using Delta Lake to have a unified view of your market data Leveraging MLflow as a delivery vehicle for model development and deployment Using Apache Spark for distributing Monte Carlo simulations at scale The ability to efficiently slice and dice your Monte Carlo simulations in order to have a more agile and forward-looking approach to risk management will be covered in a second blog post, focused more on a risk analyst persona.
Modernizing data management with Delta Lake With the rise of big data and cloud based-technologies, the IT landscape has drastically changed in the last decade.
Yet, most FSIs still rely on mainframes and non-distributed databases for core risk operations such as VaR calculations and move only some of their downstream processes to modern data lakes and cloud infrastructure.
As a result, banks are falling behind the technology curve and their current risk management practices are no longer sufficient for the modern economy.
Modernizing risk management starts with the data.
Specifically, by shifting the lense in which data is viewed: not as a cost, but as an asset.
Old Approach: When data is considered as a cost, FSIs limit the capacity of risk analysts to explore “what if“ scenarios and restrict their aggregated data silos to only satisfy predefined risk strategies.
Over time, the rigidity of maintaining silos has led engineers to branch new processes and create new aggregated views on the basis of already fragile workflows in order to adapt to evolving requirements.
Paradoxically, the constant struggle to keep data as a low cost commodity on-premises has led to a more fragile and therefore more expensive ecosystem to maintain overall.
Failed processes (annotated as X symbol below) have far too many downstream impacts in order to guarantee both timeliness and reliability of your data.
Consequently, having an intra-day (and reliable) view of market risk has become increasingly complex and cost prohibitive to achieve given all the moving components and inter-dependencies as schematised in below diagram.
Modern Approach: When data is considered as an asset, organizations embrace the versatile nature of the data, serving multiple use cases (such as value-at-risk and expected shortfall) and enabling a variety of ad-hoc analysis (such as understanding risk exposure to a specific country).
Risk analysts are no longer restricted to a narrow view of the risk and can adopt a more agile approach to risk management.
By unifying streaming and batch ETL, ensuring ACID compliance and schema enforcement, Delta Lake brings performance and reliability to your data lake, gradually increasing the quality and relevance of your data through its bronze, silver and gold layers and bridging the gap between operation processes and analytics data.
In this demo, we evaluate the level of risk of various investments in a Latin America equity portfolio composed of 40 instruments across multiple industries, storing all returns in a centralized Delta Lake table that will drive all our value-at-risk calculations (covered in our part 2 demo).
For the purpose of this demo, we access daily close prices from Yahoo finance using python yfinance library.
In real life, one may acquire market data from source systems directly (such as change data capture from mainframes) to a Delta Lake table, storing raw information on Bronze and curated / validated data on a Silver table, in real-time.
With our core data available on Delta Lake, we apply a simple window function to compute daily log returns and output results back to a gold table ready for risk modelling and analysis.
@udf(“double”) def compute_return(first, close): return float(np.
log(close / first)) window = Window.
rowsBetween(-1, 0) spark .
withColumn(“return”, compute_return(first, close)) .
select(date, ticker, return) .
saveAsTable(stock_data_gold) In the example below, we show a specific slice of our investment data for AVAL (Grupo Aval Acciones y Valores S.
A), a financial services company operating in Columbia.
Given the expected drop in its stock price post march 2020, we can evaluate its impact on our overall risk portfolio.
Streamlining model development with MLFlow Although quantitative analysis is not a new concept, the recent rise of data science and the explosion of data volumes has uncovered major inefficiencies in the way banks operate models.
Without any industry standard, data scientists often operate on a best effort basis.
This often means training models against data samples on single nodes and manually tracking models throughout the development process, resulting in long release cycles (it may take between 6 to 12 months to deliver a model to production).
The long model development cycle hinders the ability for them to quickly adapt to emerging threats and to dynamically mitigate the associated risks.
The major challenge FSIs face in this paradigm is reducing model development-to-production time without doing so at the expense of governance and regulations or contributing to an even more fragile data science ecosystem.
MLflow is the de facto standard for managing the machine learning lifecycle by bringing immutability and transparency to model development, but is not restricted to AI.
A bank’s definition of a model is usually quite broad and includes any financial models from Excel macros to rule-based systems or state-of-the art machine learning, all of them that could benefit from having a central model registry provided by MLflow within Databricks Unified Data Analytics Platform.
Reproducing model development In this example, we want to train a new model that predicts stock returns given market indicators (such as S&P 500, crude oil and treasury bonds).
We can retrieve “AS OF“ data in order to ensure full model reproducibility and audit compliance.
This capability of Delta Lake is commonly referred to as “time travel“.
The resulting data set will remain consistent throughout all experiments and can be accessed as-is for audit purposes.
DESCRIBE HISTORY market_return; SELECT * FROM market_return TIMESTAMP AS OF 2020-05-04; SELECT * FROM market_return VERSION AS OF 2; In order to select the right features in their models, quantitative analysts often navigate between Spark and Pandas dataframes.
We show here how to switch from a pyspark to python context in order to extract correlations of our market factors.
The Databricks interactive notebooks come with built-in visualisations and also fully support the use of Matplotlib, seaborn (or ggplot2 for R).
factor_returns_pd = factor_returns_df.
toPandas() factor_corr = factor_returns_pd.
corr(method=spearman, min_periods=12) Assuming our indicators are not correlated (they are) and predictive of our portfolio returns (they may), we want to log this graph as evidence to our successful experiment.
This shows internal audit, model validation functions as well as regulators that model exploration was conducted with highest quality standards and its development was led with empirical results.
png) Training models in parallel As the number of instruments in our portfolio increases, we may want to train models in parallel.
This can be achieved through a simple Pandas UDF function as follows.
For convenience (models may be more complex in real life), we want to train a simple linear regression model and aggregate all model coefficients as a n x m matrix (n being the number of instruments and m the number of features derived from our market factors).
schema = StructType([ StructField(ticker, StringType(), True), StructField(weights, ArrayType(FloatType()), True) ]) @pandas_udf(schema, PandasUDFType.
GROUPED_MAP) def train_model(group, pdf): X = np.
array(pdf[features]) X = sm.
add_constant(X, prepend=True) y = np.
array(pdf[return]) model = sm.
fit() w_df = pd.
params]], columns=[weights]) w_df[ticker] = group return w_df models_df = x_train.
toPandas() The resulting dataset (weight for each model) can be easily collected back to memory and logged to MLflow as our model candidate for the rest of the experiment.
In the below graph, we report the predicted vs actual stock return derived from our model for Ecopetrol S.
, an oil and gas producer in Columbia.
Our experiment is now stored on MLflow alongside all evidence required for an independent validation unit (IVU) submission which is likely a part of your model risk management framework.
It is key to note that this experiment is not only linked to our notebook, but to the exact revision of it, bringing independent experts and regulators the full traceability of our model as well all the necessary context required for model validation.
Monte Carlo simulations at scale with Apache Spark Value-at-risk is the process of simulating random walks that cover possible outcomes as well as worst case (n) scenarios.
A 95% value-at-risk for a period of (t) days is the best case scenario out of the worst 5% trials.
We therefore want to generate enough simulations to cover a range of possible outcomes given a 90 days historical market volatility observed across all the instruments in our portfolio.
Given the number of simulations required for each instrument, this system must be designed with a high degree of parallelism in mind, making value-at-risk the perfect workload to execute in a cloud based environment.
Risk management is the number one reason top tier banks evaluate cloud compute for analytics today and accelerate value through the Databricks runtime.
Creating a multivariate distribution Whilst the industry recommends generating between 20 to 30 thousands simulations, the main complexity of calculating value-at-risk for a mixed portfolio is not to measure individual assets returns, but the correlations between them.
At a portfolio level, market indicators can be elegantly manipulated within native python without having to shift complex matrix computation to a distributed framework.
As it is common to operate with multiple books and portfolios, this same process can easily scale out by distributing matrix calculation in parallel.
We use the last 90 days of market returns in order to compute todays’ volatility (extracting both average and covariance).
def retrieve_market_factors(from_date, to_date): from_ts = F.
cast(TimestampType()) to_ts = F.
cast(TimestampType()) f_ret = spark.
col(date) > from_ts) .