Showing 5 of total 5 results (show query)
citoverse
cito:Building and Training Neural Networks
The 'cito' package provides a user-friendly interface for training and interpreting deep neural networks (DNN). 'cito' simplifies the fitting of DNNs by supporting the familiar formula syntax, hyperparameter tuning under cross-validation, and helps to detect and handle convergence problems. DNNs can be trained on CPU, GPU and MacOS GPUs. In addition, 'cito' has many downstream functionalities such as various explainable AI (xAI) metrics (e.g. variable importance, partial dependence plots, accumulated local effect plots, and effect estimates) to interpret trained DNNs. 'cito' optionally provides confidence intervals (and p-values) for all xAI metrics and predictions. At the same time, 'cito' is computationally efficient because it is based on the deep learning framework 'torch'. The 'torch' package is native to R, so no Python installation or other API is required for this package.
Maintained by Maximilian Pichler. Last updated 3 months ago.
machine-learningneural-network
42 stars 9.07 score 129 scripts 1 dependentsnredell
forecastML:Time Series Forecasting with Machine Learning Methods
The purpose of 'forecastML' is to simplify the process of multi-step-ahead forecasting with standard machine learning algorithms. 'forecastML' supports lagged, dynamic, static, and grouping features for modeling single and grouped numeric or factor/sequence time series. In addition, simple wrapper functions are used to support model-building with most R packages. This approach to forecasting is inspired by Bergmeir, Hyndman, and Koo's (2018) paper "A note on the validity of cross-validation for evaluating autoregressive time series prediction" <doi:10.1016/j.csda.2017.11.003>.
Maintained by Nickalus Redell. Last updated 5 years ago.
deep-learningdirect-forecastingforecastforecastingmachine-learningmulti-step-ahead-forecastingneural-networkpythontime-series
130 stars 7.64 score 134 scriptsmidasverse
rMIDAS:Multiple Imputation with Denoising Autoencoders
A tool for multiply imputing missing data using 'MIDAS', a deep learning method based on denoising autoencoder neural networks. This algorithm offers significant accuracy and efficiency advantages over other multiple imputation strategies, particularly when applied to large datasets with complex features. Alongside interfacing with 'Python' to run the core algorithm, this package contains functions for processing data before and after model training, running imputation model diagnostics, generating multiple completed datasets, and estimating regression models on these datasets.
Maintained by Thomas Robinson. Last updated 1 years ago.
deep-learningimputation-methodsneural-networkreticulatetensorflow
34 stars 6.53 score 33 scriptscmusso86
recalibratiNN:Quantile Recalibration for Regression Models
Enables the diagnostics and enhancement of regression model calibration.It offers both global and local visualization tools for calibration diagnostics and provides one recalibration method: Torres R, Nott DJ, Sisson SA, Rodrigues T, Reis JG, Rodrigues GS (2024) <doi:10.48550/arXiv.2403.05756>. The method leverages on Probabilistic Integral Transform (PIT) values to both evaluate and perform the calibration of statistical models. For a more detailed description of the package, please refer to the bachelor's thesis available bellow.
Maintained by Carolina Musso. Last updated 2 months ago.
calibrationgaussian-modelsneural-networkprobabilityrecalibrationregression-models
7 stars 5.32 score 8 scriptsdiegommcc
SpatialDDLS:Deconvolution of Spatial Transcriptomics Data Based on Neural Networks
Deconvolution of spatial transcriptomics data based on neural networks and single-cell RNA-seq data. SpatialDDLS implements a workflow to create neural network models able to make accurate estimates of cell composition of spots from spatial transcriptomics data using deep learning and the meaningful information provided by single-cell RNA-seq data. See Torroja and Sanchez-Cabo (2019) <doi:10.3389/fgene.2019.00978> and Mañanes et al. (2024) <doi:10.1093/bioinformatics/btae072> to get an overview of the method and see some examples of its performance.
Maintained by Diego Mañanes. Last updated 5 months ago.
deconvolutiondeep-learningneural-networkspatial-transcriptomics
5 stars 4.88 score 1 scripts