Sanjiv Das’s research while affiliated with Santa Clara University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (19)


Banking networks, systemic risk, and the credit cycle in emerging markets
  • Article

August 2022

·

24 Reads

·

8 Citations

Journal of International Financial Markets Institutions and Money

Sanjiv R. Das

·

·

Subhankar Nayak

We study how globalization impacts systemic risk in emerging markets. We extend a large literature on systemic risk in the US, Europe, and other developed countries to emerging markets, which are relatively under-researched. Our findings are based on a large-scale empirical examination of systemic risk among 1048 financial institutions in a sample of 23 emerging markets, broken down into 5 regions, along with 369 U.S. financial institutions. Using an additively decomposable systemic risk score that combines banking system interconnectedness with default probabilities, systemic risk is quantified for each region, across time. The empirical analyses suggest that emerging markets’ systemic risk is heterogeneous across regions, is strongly dependent on the interconnectedness of the banking system within each region, and drives the level of default risk in each region, while the regions are compartmentalized away from each other and insulated from the United States. The systemic risk score may be used as a policy variable in each emerging market region to manage the credit cycle. Our evidence is consistent with the notion that globalization engenders financial stability and does not lead to large systemic risk spillovers across emerging market regions.


Digitization and Data Frames for Card Index Records

July 2022

·

7 Reads

·

3 Citations

Explorations in Economic History

We develop a methodology for converting card index archival records into usable data frames for statistical and textual analyses. Leveraging machine learning and natural-language processing tools from Amazon Web Services (AWS), we overcome hurdles associated with character recognition, inconsistent data reporting, column misalignment, and irregular naming. In this article, we detail the step-by-step conversion process and discuss remedies for common problems and edge cases, using historical records from the Reconstruction Finance Corporation.


Figure 7: Implausibility measures 1 (left column) and 2 (right column) versus failures rates for different sets of hyperparameters of the ICE and DPE algorithms and their sparse variants applied to the NCAD (first row) and USAD (second row) models on a the KPI dataset. The metrics are computed over a validation set of 5 time series and the failure rate's threshold is 10% (red dotted line).
Figure 8: Implausibility measures 1 (left column) and 2 (right column) versus failures rates for different sets of hyperparameters of the ICE and DPE algorithms and their sparse variants applied to the NCAD (first row) and USAD (second row) models on a the Yahoo dataset. The metrics are computed over a validation set of 15 time series and the failure rate's threshold is 25% (red dotted line).
Figure 10: Implausibility measures 1 (left column) and 2 (right column) versus failures rates for different sets of hyperparameters of the ICE and DPE algorithms and their sparse variants applied to the NCAD (first row) and USAD (second row) models on a the SMD dataset. The metrics are computed over a validation set of 6 time series and the failure rate's threshold is 40% for NCAD and 20% for USAD (red dotted lines).
Figure 11: Diversity of the counterfactual ensemble (left) and failure rate of our counterfactual method (right) versus the learning rate of the SGD algorithm for the two variants of our method, ICE and DPE.
Hyperparameters of the Sparse ICE algorithm on the two benchmark multivariate datasets.
Diverse Counterfactual Explanations for Anomaly Detection in Time Series
  • Preprint
  • File available

March 2022

·

98 Reads

·

1 Citation

Deborah Sulem

·

·

Muhammad Bilal Zafar

·

[...]

·

Cedric Archambeau

Data-driven methods that detect anomalies in times series data are ubiquitous in practice, but they are in general unable to provide helpful explanations for the predictions they make. In this work we propose a model-agnostic algorithm that generates counterfactual ensemble explanations for time series anomaly detection models. Our method generates a set of diverse counterfactual examples, i.e, multiple perturbed versions of the original time series that are not considered anomalous by the detection model. Since the magnitude of the perturbations is limited, these counterfactuals represent an ensemble of inputs similar to the original time series that the model would deem normal. Our algorithm is applicable to any differentiable anomaly detection model. We investigate the value of our method on univariate and multivariate real-world datasets and two deep-learning-based anomaly detection models, under several explainability criteria previously proposed in other data domains such as Validity, Plausibility, Closeness and Diversity. We show that our algorithm can produce ensembles of counterfactual examples that satisfy these criteria and thanks to a novel type of visualisation, can convey a richer interpretation of a model's internal mechanism than existing methods. Moreover, we design a sparse variant of our method to improve the interpretability of counterfactual explanations for high-dimensional time series anomalies. In this setting, our explanation is localised on only a few dimensions and can therefore be communicated more efficiently to the model's user.

Download

Figure 3: Examples of the landing page and the highlighted text from the Wikipedia human surveys.
Figure 4: Examples of the landing page and the highlighted text from the IMDB human surveys.
Detailed statistics of the datasets used in the experiments. The columns Words and Sentences show the average ± standard deviation across the data. Prev. Most shows the prevalence (in percentage) of the most prevalent class in the dataset.
Percentage of common predictions between different initializations.
More Than Words: Towards Better Quality Interpretations of Text Classifiers

December 2021

·

28 Reads

·

1 Citation

The large size and complex decision mechanisms of state-of-the-art text classifiers make it difficult for humans to understand their predictions, leading to a potential lack of trust by the users. These issues have led to the adoption of methods like SHAP and Integrated Gradients to explain classification decisions by assigning importance scores to input tokens. However, prior work, using different randomization tests, has shown that interpretations generated by these methods may not be robust. For instance, models making the same predictions on the test set may still lead to different feature importance rankings. In order to address the lack of robustness of token-based interpretability, we explore explanations at higher semantic levels like sentences. We use computational metrics and human subject studies to compare the quality of sentence-based interpretations against token-based ones. Our experiments show that higher-level feature attributions offer several advantages: 1) they are more robust as measured by the randomization tests, 2) they lead to lower variability when using approximation-based methods like SHAP, and 3) they are more intelligible to humans in situations where the linguistic coherence resides at a higher granularity level. Based on these findings, we show that token-based interpretability, while being a convenient first choice given the input interfaces of the ML models, is not the most effective one in all situations.


Bank Regulation, Network Topology, and Systemic Risk: Evidence from the Great Depression

December 2021

·

70 Reads

·

8 Citations

Journal of Money Credit and Banking

We study how bank regulation interacts with network topology to influence systemic stability. Employing unique hand‐collected data on the correspondent network for all U.S. banks prior to the Great Depression and a methodology that captures bank credit risk and network position, we demonstrate how the pyramid‐shaped network topology was inherently fragile and systemically risky. We measure its contribution to banking distress, and show that a bank's network position as well as its network neighbors' risk are strong predictors of bank survivorship. Institutional alternatives, such as branch banking, and alternative topologies deliver networks that are more stable than that of 1929.


FinLex: An Effective Use of Word Embeddings for Financial Lexicon Generation

October 2021

·

48 Reads

·

13 Citations

The Journal of Finance and Data Science

We present a simple and effective methodology for the generation of lexicons (word lists) that may be used in natural language scoring applications. In particular, in the finance industry, word lists have become ubiquitous for sentiment scoring. These have been derived from dictionaries such as the Harvard Inquirer and require manual curation. Here, we present an automated approach to the curation of lexicons, which makes automatic preparation of any word list immediate. We show that our automated word lists deliver comparable performance to traditional lexicons on machine learning classification tasks. This new approach will enable finance academics and practitioners to create and deploy new word lists in addition to the few traditional ones in a facile manner.



Amazon SageMaker Clarify: Machine Learning Bias Detection and Explainability in the Cloud

September 2021

·

311 Reads

Understanding the predictions made by machine learning (ML) models and their potential biases remains a challenging and labor-intensive task that depends on the application, the dataset, and the specific model. We present Amazon SageMaker Clarify, an explainability feature for Amazon SageMaker that launched in December 2020, providing insights into data and ML models by identifying biases and explaining predictions. It is deeply integrated into Amazon SageMaker, a fully managed service that enables data scientists and developers to build, train, and deploy ML models at any scale. Clarify supports bias detection and feature importance computation across the ML lifecycle, during data preparation, model evaluation, and post-deployment monitoring. We outline the desiderata derived from customer input, the modular architecture, and the methodology for bias and explanation computations. Further, we describe the technical challenges encountered and the tradeoffs we had to make. For illustration, we discuss two customer use cases. We present our deployment results including qualitative customer feedback and a quantitative evaluation. Finally, we summarize lessons learned, and discuss best practices for the successful adoption of fairness and explanation tools in practice.



Multimodal Machine Learning for Credit Modeling

July 2021

·

42 Reads

·

11 Citations

Credit ratings are traditionally generated using models that use financial statement data and market data, which is tabular (numeric and categorical). Practitioner and academic models do not include text data. Using an automated approach to combine long-form text from SEC filings with the tabular data, we show how multimodal machine learning using stack ensembling and bagging can generate more accurate rating predictions. This paper demonstrates a methodology to use big data to extend tabular data models, which have been used by the ratings industry for decades, to the class of multimodal machine learning models.


Citations (16)


... Economic theory suggests that systemic risk accumulates during the expansion phase of the financial cycle and materialises into financial crises during downturns (Borio et al., 2020a;Das et al., 2022;Danthine, 2012). During expansions, financial agents often become overly optimistic, leading to increased borrowing, lending, and investment in riskier assets, which appear less dangerous in a booming economy. ...

Reference:

The Stability of the Financial Cycle: Insights from a Markov Switching Regression in South Africa
Banking networks, systemic risk, and the credit cycle in emerging markets
  • Citing Article
  • August 2022

Journal of International Financial Markets Institutions and Money

... Les taules han estat transcrites automàticament mitjançant la intel·ligència artificial i l'extractor de text AWS Textract, que utilitza la computació en línia i algoritmes predefinits OCR o Optical Character Recognition per a l'aprenentatge automàtic i el reconeixement de caràcters escrits en fons documentals de diversa naturalesa (Amujala et al., 2023;Correia i Luck, 2023). Textract és especialment sensible a la transcripció de documents històrics que no poden ser transcrits de manera automàtica o semiautomàtica amb altres eines ofimàtiques més habituals, com és el cas de les còpies digitals dels oficis mecanografiats del CTV (Figura 3). ...

Digitization and Data Frames for Card Index Records
  • Citing Article
  • July 2022

Explorations in Economic History

... In their work, D. Sulem et al. [5] proposed a new method that explains the anomalies found in time series by generating counterfactual explanations. Counterfactual explanations are alternative scenarios that show how the data must change to remove an anomalous observation. ...

Diverse Counterfactual Explanations for Anomaly Detection in Time Series

... Here, we focus on global reconstructions of BERT's predictions for token-level classifications in this work, since this constitutes popular application scenarios of BERT (e.g., AS1, AS3) and since BERT also establishes text representations based on tokens. Moreover, as Zafar et al. (2021) and Yan et al. (2022) indicate, a reconstruction approach for token-level classifications can also serve as a basis for reconstructions of coarser classification tasks, for instance, for sentence-level classifications (e.g., AS2, AS4). ...

More Than Words: Towards Better Quality Interpretations of Text Classifiers

... The covariates in x i1 are measured in 1929 and include controls for balance sheet ratios, bank size (log of total assets), town and county characteristics, regulatory variables, market share of deposits, interbank network controls, and Federal Reserve district indicators. The variables selected for the bank survival equation align closely with existing work on the Great Depression (Calomiris and Mason, 2003;Das et al., 2022). A full list of the variables and their definitions is provided in Section 4.1. ...

Bank Regulation, Network Topology, and Systemic Risk: Evidence from the Great Depression
  • Citing Article
  • December 2021

Journal of Money Credit and Banking

... Since Word2Vec was introduced by Mikolov et al. (2013), studies in economics and finance that adopt this method to explore financial documents have gained in popularity; see Das et al. (2022), Li et al. (2021), Ma et al. (2023), and Miranda-Belmonte et al. (2023), among others. The ability to capture the immediate context when representing words is the key feature that sets Word2Vec apart from count-based word representation methods, which have been widely used in economic research using textual data (Henry & Leone, 2016;A.H. Huang, Zang, and Zheng, 2014;Jegadeesh & Wu, 2013;Jiang, Lee, Martin, and Zhou, 2019;Loughran & McDonald, 2011). ...

FinLex: An Effective Use of Word Embeddings for Financial Lexicon Generation
  • Citing Article
  • October 2021

The Journal of Finance and Data Science

... While these metrics are widely used, they can exhibit biases related to sample size. As such, this bias affects common fairness metrics, including disparate impact (Feldman et al., 2015), equalized odds (Hardt et al., 2016), and predictive parity (Das et al., 2021). By highlighting this issue, we aim to enhance the robustness of metric comparisons across datasets with varying sample sizes. ...

Fairness Measures for Machine Learning in Finance
  • Citing Article
  • September 2021

The Journal of Financial Data Science

... While a number of studies have begun to explore this textual modality, it's fair to say that it has not been fully capitalized on in CCR analysis. These studies typically rely on traditional feature extraction techniques like sentiment analysis, N-gram models, bag-of-words, and document embeddings (e.g., doc2vec), which may not capture semantic and syntactic information as effectively as modern deep learning approaches (Choi et al., 2020;Nguyen et al., 2021;Tsai & Wang, 2017;Wang et al., 2023). In terms of multimodal fusion, the prevalent approaches involve simply concatenating features from different modalities without considering the inherent interconnections between them (Mai et al., 2019;Stevenson et al., 2021). ...

Multimodal Machine Learning for Credit Modeling
  • Citing Conference Paper
  • July 2021

... Google's PAIR team, for example, released the What-If Tool [32] and Fairness Indicators [33], which allow developers to visualize model behavior for different slices of data and compute basic bias metrics. There are also fairness libraries like Themis [34]or AEC (Audit AI) [35] and industry services such as Amazon SageMaker Clarify [36]. SageMaker Clarify is a service that helps detect bias in machine learning data and models; it can analyze datasets for bias by requiring the user to specify which features are sensitive (like gender or age) and then computes bias metrics and produces a report. ...

Amazon SageMaker Clarify: Machine Learning Bias Detection and Explainability in the Cloud
  • Citing Conference Paper
  • August 2021