You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Implementation of diffusion models with varying noise distributions (Gaussian, GMM, Gamma) and scheduling techniques (cosine, sigmoid) to assess generative performance using KL divergence and dynamic scheduling approaches.
Using Monte-Carlo simulated datasets, a completely transparent Boltzmann Machine trained on 1-D Ising chain data is implemented to predict model couplers in the absence of past coupler values. Methods from machine learning applied to theoretical physics are on display in this work.
Implementation of the Non-negative Multiple Matrix Factorization (NMMF) algorithm proposed in Takeuchi et al, 2013 with some modifications. There is a python native version NMMFlexPy and a R wrapper NMMFlexR
average-KL-divergence-calculator.py is a Python script that calculates the average KL divergence for each FASTA file in a directory and produces separate output files and a combined output file with the results.
Novel technique to fit a target distribution with a class of distributions using SVI (via NumPyro). Unlike standard SVI, our "data" is a distribution rather than a finite collection of samples.