-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
issue run_trisk #525
Comments
Dear @bergalli , I am now currently trying to run the trisk function adapted to my values, but could you please explain the meaning of the variables net_profit_margin_rawdata_ratio, debt_equity_ratio and volatility_rawdata_ratio in the _rawdata_ratiotests/testthat/test_data/ST_INPUTS_DEV/prewrangled_financial_data_stress_test.csv file Why are these variables categorical and in general is there a variables dictionary that I have possibly missed? Maany thanks in advance for your answer! |
Hi @Vladlenman, the variables you mention are remnants of the pre-processing applied on the financial data. They don't have any effect on the model, their usage is just to run data quality checks later on. More details about the preprocessing: Due to an incomplete matching of company's financial indicators to company's productions, the preprocessing fills in values for missing companies by using averages over sector/country. Those variables then give the information of wether the value are inferred or not. This is where the columns are created (in another repo) : https://github.com/2DegreesInvesting/STDataMGMT/blob/7915cc2aa0df1e7b5b5daaf21b9daaf11803f5a6/R/prepare_prewrangled_financial_data_stress_test.R#L494 Also for information : this repository will be archived after the summer, and the work on Trisk will be pursued here instead: Best, |
Ddear @bergalli , Thank you for your quick response. I greatly appreciate it, as it has helped me move forward. I apologize for inundating you with questions, but I am currently navigating through the repositories and trying to understand the connections between the files. Currently, I am working with input files located in tests/testthat/test_data/ST_INPUTS_DEV. Specifically, I am placing my data into abcd_input_test and prewrangled_financial_data. Everything works well as long as I use the "Global" scenarios in the scenario_geography column of the abcd_file. However, when I attempt to change the geography, I encounter the following error: -- Validating input arguments. I always verify whether the scenario is accessible for the specified region using the get_scenario_geography_x_ald_sector function. I have ensured that the combinations between geography and scenario exist in the file. I am running trisk function using the following parameters: Additionally I have a small question about the pd calculation for tests/testthat/test_data/ST_INPUTS_DEV/prewrangled_financial_data_stress_test.csv The PDs there are usually close to 0.5. I understand that you get this data from Eikon, I do not have access to it, but do really this companies have probability of default equal 50% or is this value already in percentages, so 0.5 in the file actually means 0.5%? I would appreciate any help with the following issues Best regards, |
Hi @Vladlenman , No worries it's always nice to see interest in the project :) I think your issue stems from the fact that you haven't set the input and output paths as parameters, which is where the input data is stored, and an existing output directory where the results will be written. I made this notebook for another project, which replicates the Trisk methodology end to end using the synthetic data you already saw. That should help you setup , let me know if you have issues accessing it. https://colab.research.google.com/drive/1mVFSQxOVMoIE-t5GK2StJPkcSD0_OwqK?usp=sharing |
Dear @bergalli, many thanks for the shared collab file. It has helped me to repeat the whole process and got the desired plots. I have left only a few theoretical and data management questions about the process, but before it I would clarify some technical issues that I observed and would be very thankful about your answers. -Unfortunately, I did not manage to run the function for any geographical scenario except for global, and even for the synthetic dataset presented in this git it works only for Global scenario. Did I get it right that currently it is only possible to run the function for the global scenario? Other questions are more related to the variable description interpretation of the results: And if you have some more energy, I would be interested in you some future development plans: Many thanks in advance for your help. I enjoy engaging with your project! |
devtools::load_all()
The text was updated successfully, but these errors were encountered: