Showing 17 of total 17 results (show query)
mhahsler
arules:Mining Association Rules and Frequent Itemsets
Provides the infrastructure for representing, manipulating and analyzing transaction data and patterns (frequent itemsets and association rules). Also provides C implementations of the association mining algorithms Apriori and Eclat. Hahsler, Gruen and Hornik (2005) <doi:10.18637/jss.v014.i15>.
Maintained by Michael Hahsler. Last updated 2 months ago.
arulesassociation-rulesfrequent-itemsets
194 stars 13.99 score 3.3k scripts 28 dependentsbioc
IPO:Automated Optimization of XCMS Data Processing parameters
The outcome of XCMS data processing strongly depends on the parameter settings. IPO (`Isotopologue Parameter Optimization`) is a parameter optimization tool that is applicable for different kinds of samples and liquid chromatography coupled to high resolution mass spectrometry devices, fast and free of labeling steps. IPO uses natural, stable 13C isotopes to calculate a peak picking score. Retention time correction is optimized by minimizing the relative retention time differences within features and grouping parameters are optimized by maximizing the number of features showing exactly one peak from each injection of a pooled sample. The different parameter settings are achieved by design of experiment. The resulting scores are evaluated using response surface models.
Maintained by Thomas Lieb. Last updated 5 months ago.
immunooncologymetabolomicsmassspectrometry
34 stars 8.14 score 41 scriptssymbolixau
googlePolylines:Encoding Coordinates into 'Google' Polylines
Encodes simple feature ('sf') objects and coordinates, and decodes polylines using the 'Google' polyline encoding algorithm (<https://developers.google.com/maps/documentation/utilities/polylinealgorithm>).
Maintained by David Cooley. Last updated 15 days ago.
geospatialgisgoogle-mapspolyline-encoderr-spatialspatialcpp
18 stars 8.11 score 9 dependentskwb-r
kwb.utils:General Utility Functions Developed at KWB
This package contains some small helper functions that aim at improving the quality of code developed at Kompetenzzentrum Wasser gGmbH (KWB).
Maintained by Hauke Sonnenberg. Last updated 1 years ago.
8 stars 7.33 score 12 scripts 78 dependentsjeroen
base64:Base64 Encoder and Decoder
Compatibility wrapper to replace the orphaned package. New applications should use base64 encoders from 'jsonlite' or 'openssl' or 'base64enc'.
Maintained by Jeroen Ooms. Last updated 6 months ago.
2 stars 6.62 score 163 scripts 42 dependentsextendr
b64:Fast and Vectorized Base 64 Engine
Provides a fast, lightweight, and vectorized base 64 engine to encode and decode character and raw vectors as well as files stored on disk. Common base 64 alphabets are supported out of the box including the standard, URL-safe, bcrypt, crypt, 'BinHex', and IMAP-modified UTF-7 alphabets. Custom engines can be created to support unique base 64 encoding and decoding needs.
Maintained by Josiah Parry. Last updated 2 months ago.
16 stars 6.03 score 4 scripts 3 dependentsbergsmat
yamlet:Versatile Curation of Table Metadata
A YAML-based mechanism for working with table metadata. Supports compact syntax for creating, modifying, viewing, exporting, importing, displaying, and plotting metadata coded as column attributes. The 'yamlet' dialect is valid 'YAML' with defaults and conventions chosen to improve readability. See ?yamlet, ?decorate, ?modify, ?io_csv, and ?ggplot.decorated.
Maintained by Tim Bergsma. Last updated 1 months ago.
2 stars 5.99 score 60 scripts 1 dependentsrapler
dst:Using the Theory of Belief Functions
Using the Theory of Belief Functions for evidence calculus. Basic probability assignments, or mass functions, can be defined on the subsets of a set of possible values and combined. A mass function can be extended to a larger frame. Marginalization, i.e. reduction to a smaller frame can also be done. These features can be combined to analyze small belief networks and take into account situations where information cannot be satisfactorily described by probability distributions.
Maintained by Peiyuan Zhu. Last updated 14 hours ago.
6 stars 5.98 score 126 scriptshenrikbengtsson
aroma.affymetrix:Analysis of Large Affymetrix Microarray Data Sets
A cross-platform R framework that facilitates processing of any number of Affymetrix microarray samples regardless of computer system. The only parameter that limits the number of chips that can be processed is the amount of available disk space. The Aroma Framework has successfully been used in studies to process tens of thousands of arrays. This package has actively been used since 2006.
Maintained by Henrik Bengtsson. Last updated 1 years ago.
infrastructureproprietaryplatformsexonarraymicroarrayonechannelguidataimportdatarepresentationpreprocessingqualitycontrolvisualizationreportwritingacghcopynumbervariantsdifferentialexpressiongeneexpressionsnptranscriptionaffymetrixanalysiscopy-numberdnaexpressionhpclarge-scalenotebookreproducibilityrna
10 stars 5.79 score 112 scripts 3 dependentsmunterfi
flexpolyline:Flexible Polyline Encoding
Binding to the C++ implementation of the flexible polyline encoding by HERE <https://github.com/heremaps/flexible-polyline>. The flexible polyline encoding is a lossy compressed representation of a list of coordinate pairs or coordinate triples. The encoding is achieved by: (1) Reducing the decimal digits of each value; (2) encoding only the offset from the previous point; (3) using variable length for each coordinate delta; and (4) using 64 URL-safe characters to display the result.
Maintained by Merlin Unterfinger. Last updated 2 years ago.
gisheremapspolylinepolyline-decoderpolyline-encoderrspatialcpp
9 stars 5.75 score 14 scripts 1 dependentsbflammers
ANN2:Artificial Neural Networks for Anomaly Detection
Training of neural networks for classification and regression tasks using mini-batch gradient descent. Special features include a function for training autoencoders, which can be used to detect anomalies, and some related plotting functions. Multiple activation functions are supported, including tanh, relu, step and ramp. For the use of the step and ramp activation functions in detecting anomalies using autoencoders, see Hawkins et al. (2002) <doi:10.1007/3-540-46145-0_17>. Furthermore, several loss functions are supported, including robust ones such as Huber and pseudo-Huber loss, as well as L1 and L2 regularization. The possible options for optimization algorithms are RMSprop, Adam and SGD with momentum. The package contains a vectorized C++ implementation that facilitates fast training through mini-batch learning.
Maintained by Bart Lammers. Last updated 4 years ago.
anomaly-detectionartificial-neural-networksautoencodersneural-networksrobust-statisticsopenblascppopenmp
13 stars 5.59 score 60 scriptsbquast
HomomorphicEncryption:BFV, BGV, CKKS Schema for Fully Homomorphic Encryption
Implements the Brakerski-Fan-Vercauteren (BFV, 2012) <https://eprint.iacr.org/2012/144>, Brakerski-Gentry-Vaikuntanathan (BGV, 2014) <doi:10.1145/2633600>, and Cheon-Kim-Kim-Song (CKKS, 2016) <https://eprint.iacr.org/2016/421.pdf> schema for Fully Homomorphic Encryption. The included vignettes demonstrate the encryption procedures.
Maintained by Bastiaan Quast. Last updated 1 years ago.
1 stars 5.52 score 39 scriptsmyominnoo
mStats:Medical Statistics & Epidemiological Analysis
A set of tidyverse-friendly functions for data management, calculation of epidemiological measures, statistical analysis, and table creation.
Maintained by Myo Minn Oo. Last updated 1 years ago.
data-managementepidemiological-calculationsmedical-statistics
4.98 score 16 scripts 1 dependentsolisansonwu
diyar:Record Linkage and Epidemiological Case Definitions in 'R'
An R package for iterative and batched record linkage, and applying epidemiological case definitions. 'diyar' can be used for deterministic and probabilistic record linkage, or multistage record linkage combining both approaches. It features the implementation of nested match criteria, and mechanisms to address missing data and conflicting matches during stepwise record linkage. Case definitions are implemented by assigning records to groups based on match criteria such as person or place, and overlapping time or duration of events e.g. sample collection dates or periods of hospital stays. Matching records are assigned a unique group ID. Index and duplicate records are removed or further analyses as required.
Maintained by Olisaeloka Nsonwu. Last updated 3 months ago.
6 stars 4.77 score 33 scriptsalshum
hashids:Generate Short Unique YouTube-Like IDs (Hashes) from Integers
An R port of the hashids library. hashids generates YouTube-like hashes from integers or vector of integers. Hashes generated from integers are relatively short, unique and non-seqential. hashids can be used to generate unique ids for URLs and hide database row numbers from the user. By default hashids will avoid generating common English cursewords by preventing certain letters being next to each other. hashids are not one-way: it is easy to encode an integer to a hashid and decode a hashid back into an integer.
Maintained by Alex Shum. Last updated 6 years ago.
18 stars 4.10 score 14 scriptsbergsmat
encode:Represent Ordered Lists and Pairs as Strings
Interconverts between ordered lists and compact string notation. Useful for capturing code lists, and pair-wise codes and decodes, for text storage. Analogous to factor levels and labels. Generics encode() and decode() perform interconversion, while codes() and decodes() extract components of an encoding. The function encoded() checks whether something is interpretable as an encoding. If a vector has an encoded 'guide' attribute, as_factor() uses it to coerce to factor.
Maintained by Tim Bergsma. Last updated 6 years ago.
2 stars 4.03 score 12 scripts 5 dependents