QuickStart - minimal configuration guide#
You can copy one of the tutorial notebooks found in this documentation and customize it for your project. The customization includes setting serveral parameters depending the type of the analysis.
π Easy Configuration with Streamlit App#
NEW: Use our interactive configuration interface!
Instead of manually setting parameters, you can use our Streamlit-based configuration interface:
Online version: https://maitconfig.streamlit.app/
Local version: Run
streamlit run streamlit_app.py
in your MAIT directory
The Streamlit app provides:
β Interactive parameter configuration with validation
β Real-time preview of your settings
β Export ready-to-run Python scripts
β Built-in help and documentation
β No manual coding required
Simply configure your pipeline through the web interface and download the generated Python script!
Manual Configuration (Alternative Method)#
General Settings (Mandatory for All Tasks)#
Specify Main Directory:
main_folder_name = 'results_<ProjectName>'
Binary Classification (Mandatory Parameters)#
Outcome Variable:
outcome_var = "target"
you must also specify the names of categorical variables (features) if available in your dataset, like cat_features = ["race", "gender"]
.
Also if you need some variables not to be included like columns_to_drop = ["irrelevant feature X", "irrelevant feature Y"]
, and also to make sure continuous variables are also defined like continuous_features = ["age","weight"]
.
Class Labels:
Positive Class:
class_1 = "malignant"
Negative Class:
class_0 = "benign"
also replace 0 and 1 here based on your dataset
class_label_dict = {0:class_0, 1:class_1}
, maybe in your dataset theyβre like -1 and 1 instead.Data Split:
Enable/Disable:
data_split = True
Training Size:
If
data_split = True
:train_size_perc = 0.8
Already Split Data (If Applicable):
already_split = False
If
True
, specify subsets:testset = mydata[mydata['subset'] == 'Test'] mydata = mydata[mydata['subset'] == 'Train']
Classification Models:
Specify at least two models:
Survival Analysis (Mandatory Parameters)#
Enable Survival Analysis:
survival_analysis = True
Time-to-Event Column:
time_to_event_column = "max_time_difference_days"
Regression (Mandatory Parameters)#
Enable Regression Analysis:
regression_analysis = True
Outcome Variable:
regression_outcome = "regression_outcome_var"
Resource Configuration#
CPU/GPU Settings:
Enable GPU (if available):
GPU_avail = True
CPUs for Training:
n_cpu_model_training = 20
CPUs for Hyperparameter Tuning:
n_cpu_for_tuning = 20
Additional Notes#
Recommended: Use the Streamlit configuration app for easier setup and automatic parameter validation
Each task (classification, survival, regression) has specific mandatory parameters. Only include the settings relevant to the task you are running.
Omitting any of the above parameters will result in errors or crashes during execution.
For further details, refer to the full pipeline manual.