Projects supporting the matlab data format.
Showing Items 1-20 of 72 on page 1 of 4: 1 2 3 4 Next

Logo 1SpectralClustering 1.2

by tbuehler - May 1, 2018, 19:26:07 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 34349 views, 7604 downloads, 0 subscriptions

About: A fast and scalable graph-based clustering algorithm based on the eigenvectors of the nonlinear 1-Laplacian.

Changes:
  • improved optimization of ncut and rcut criterion
  • optimized eigenvector initialization
  • changed default values for number of runs
  • several internal optimizations
  • made console output more informative

Logo JMLR GPML Gaussian Processes for Machine Learning Toolbox 4.1

by hn - November 27, 2017, 19:26:13 CET [ Project Homepage BibTeX Download ] 88142 views, 19680 downloads, 0 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarWhole Star
(based on 2 votes)

About: The GPML toolbox is a flexible and generic Octave/Matlab implementation of inference and prediction with Gaussian process models. The toolbox offers exact inference, approximate inference for non-Gaussian likelihoods (Laplace's Method, Expectation Propagation, Variational Bayes) as well for large datasets (FITC, VFE, KISS-GP). A wide range of covariance, likelihood, mean and hyperprior functions allows to create very complex GP models.

Changes:

Logdet-estimation functionality for grid-based approximate covariances

  • Lanczos subspace estimation

  • Chebyshef polynomial expansion

More generic infEP functionality

  • dense computations and sparse approximations using the same code

  • covering KL inference as a special cas of EP

New infKL function contributed by Emtiyaz Khan and Wu Lin

  • Conjugate-Computation Variational Inference algorithm

  • much more scalable than previous versions

Time-series covariance functions on the positive real line

  • covW (i-times integrated) Wiener process covariance

  • covOU (i-times integrated) Ornstein-Uhlenbeck process covariance (contributed by Juan Pablo Carbajal)

  • covULL underdamped linear Langevin process covariance (contributed by Robert MacKay)

  • covFBM Fractional Brownian motion covariance

New covariance functions

  • covWarp implements k(w(x),w(z)) where w is a "warping" function

  • covMatern has been extended to also accept non-integer distance parameters


Logo Accord.NET Framework 3.8.0

by cesarsouza - October 23, 2017, 20:50:27 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 79537 views, 16036 downloads, 0 subscriptions

About: The Accord.NET Framework is a .NET machine learning framework combined with audio and image processing libraries completely written in C#. It is a complete framework for building production-grade computer vision, computer audition, signal processing and statistics applications even for commercial use. A comprehensive set of sample applications provide a fast start to get up and running quickly, and an extensive online documentation helps fill in the details.

Changes:

For a complete list of changes, please see the full release notes at the release details page at:

https://github.com/accord-net/framework/releases/tag/v3.8.0


Logo pSpectralClustering 1.2

by tbuehler - July 30, 2017, 20:07:52 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 25764 views, 5151 downloads, 0 subscriptions

About: A generalized version of spectral clustering using the graph p-Laplacian.

Changes:

various internal optimizations


Logo JMLR MSVMpack 1.5.1

by lauerfab - March 9, 2017, 12:29:37 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 64195 views, 15554 downloads, 0 subscriptions

About: MSVMpack is a Multi-class Support Vector Machine (M-SVM) package. It is dedicated to SVMs which can handle more than two classes without relying on decomposition methods and implements the four M-SVM models from the literature: Weston and Watkins M-SVM, Crammer and Singer M-SVM, Lee, Lin and Wahba M-SVM, and the M-SVM2 of Guermeur and Monfrini.

Changes:
  • Fix compilation error with recent gcc

Logo MIToolbox 3.0.1

by apocock - March 2, 2017, 00:38:52 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 65531 views, 10876 downloads, 0 subscriptions

About: A mutual information library for C and Mex bindings for MATLAB. Aimed at feature selection, and provides simple methods to calculate mutual information, conditional mutual information, entropy, conditional entropy, Renyi entropy/mutual information, and weighted variants of Shannon entropies/mutual informations. Works with discrete distributions, and expects column vectors of features.

Changes:

Fixed a Windows compilation bug. MIToolbox v3 should now compile using Visual Studio.


Logo NaN toolbox 3.1.2

by schloegl - January 22, 2017, 12:24:59 CET [ Project Homepage BibTeX Download ] 138543 views, 32070 downloads, 0 subscriptions

About: NaN-toolbox is a statistics and machine learning toolbox for handling data with and without missing values.

Changes:

Changes in v.3.1.2 - improve configuration and build system - support of more platforms (including Octave 4.2.0) improved

Changes in v.3.0.3 - improve compatibility for Octave on Windows

Changes in v.3.0.1 - fix packaging for octave

Changes in v.2.8.5 - bug fix: trimmean - compiler support for gcc-5 and clang - fix typos

For details see the CHANGELOG at http://pub.ist.ac.at/~schloegl/matlab/NaN/CHANGELOG


Logo FEAST 2.0.0

by apocock - January 8, 2017, 00:49:19 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 75881 views, 12796 downloads, 0 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarWhole Star
(based on 2 votes)

About: FEAST provides implementations of common mutual information based filter feature selection algorithms (mim, mifs, mrmr, cmim, icap, jmi, disr, fcbf, etc), and an implementation of RELIEF. Written for C/C++ & Matlab.

Changes:

Major refactoring of FEAST to improve speed and portability.

  • FEAST now clones the input data if it's floating point and discretises it to unsigned ints once in a single pass. This improves the speed by about 30%.
  • FEAST now has unsigned int entry points which avoid this discretisation and are much faster if the data is already categorical.
  • Added weighted feature selection algorithms to FEAST which can be used for cost-sensitive feature selection.
  • Added a Java API using JNI.
  • FEAST now returns the internal score for each feature according to the criterion. Available in all three APIs.
  • Rearranged the repository to make it easier to work with. Header files are now in `include`, source in `src`, the MATLAB API is in `matlab/` and the Java API is in `java/`.
  • FEAST now compiles cleanly using `-std=c89 -Wall -Werror`.

Logo slim for matlab 0.2

by ustunb - August 23, 2016, 20:27:00 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 9666 views, 2372 downloads, 0 subscriptions

About: learn optimized scoring systems using MATLAB and the CPLEX Optimization Studio

Changes:

Initial Announcement on mloss.org.


Logo Social Impact theory based Optimizer library 1.1

by rishem - July 29, 2016, 13:19:47 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 27681 views, 6058 downloads, 0 subscriptions

About: This is an optimization library based on Social Impact Theory(SITO). The optimizer works in the same way as PSO and GA.

Changes:

bug removed


Logo DeeBNet, a new object oriented MATLAB toolbox for Deep Belief Networks 3.2

by keyvanrad - June 26, 2016, 16:19:55 CET [ Project Homepage BibTeX Download ] 36566 views, 8478 downloads, 0 subscriptions

About: Nowadays this is very popular to use the deep architectures in machine learning. Deep Belief Networks (DBNs) are deep architectures that use a stack of Restricted Boltzmann Machines (RBM) to create a powerful generative model using training data. DBNs have many abilities such as feature extraction and classification that are used in many applications including image processing, speech processing, text categorization, etc. This paper introduces a new object oriented toolbox with the most important abilities needed for the implementation of DBNs. According to the results of the experiments conducted on the MNIST (image), ISOLET (speech), and the 20 Newsgroups (text) datasets, it was shown that the toolbox can learn automatically a good representation of the input from unlabeled data with better discrimination between different classes. Also on all the aforementioned datasets, the obtained classification errors are comparable to those of the state of the art classifiers. In addition, the toolbox supports different sampling methods (e.g. Gibbs, CD, PCD and our new FEPCD method), different sparsity methods (quadratic, rate distortion and our new normal method), different RBM types (generative and discriminative), GPU based, etc. The toolbox is a user-friendly open source software in MATLAB and Octave and is freely available on the website.

Changes:

New in toolbox

  • Using GPU in Backpropagation
  • Revision of some demo scripts
  • Function approximation with multiple outputs
  • Feature extraction with GRBM in first layer

cardinal


Logo JMLR Information Theoretical Estimators 0.63

by szzoli - June 9, 2016, 23:42:14 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 325709 views, 61440 downloads, 0 subscriptions

About: ITE (Information Theoretical Estimators) is capable of estimating many different variants of entropy, mutual information, divergence, association measures, cross quantities and kernels on distributions. Thanks to its highly modular design, ITE supports additionally (i) the combinations of the estimation techniques, (ii) the easy construction and embedding of novel information theoretical estimators, and (iii) their immediate application in information theoretical optimization problems.

Changes:
  • Conditional Shannon entropy estimation: added.

  • Conditional Shannon mutual information estimation: included.


Logo JMLR GPstuff 4.7

by avehtari - June 9, 2016, 17:45:15 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 91186 views, 20855 downloads, 0 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarWhole Star
(based on 1 vote)

About: The GPstuff toolbox is a versatile collection of Gaussian process models and computational tools required for inference. The tools include, among others, various inference methods, sparse approximations and model assessment methods.

Changes:

2016-06-09 Version 4.7

Development and release branches available at https://github.com/gpstuff-dev/gpstuff

New features

  • Simple Bayesian Optimization demo

Improvements

  • Improved use of PSIS
  • More options added to gp_monotonic
  • Monotonicity now works for additive covariance functions with selected variables
  • Possibility to use gpcf_squared.m-covariance function with derivative observations/monotonicity
  • Default behaviour made more robust by changing default jitter from 1e-9 to 1e-6
  • LA-LOO uses the cavity method as the default (see Vehtari et al (2016). Bayesian leave-one-out cross-validation approximations for Gaussian latent variable models. JMLR, accpeted for publication)
  • Selected variables -option works now better with monotonicity

Bugfixes

  • small error in derivative observation computation fixed
  • several minor bug fixes

Logo MDLText 1

by renatoms88 - March 3, 2016, 19:31:25 CET [ BibTeX Download ] 4605 views, 1659 downloads, 0 subscriptions

About: testing mloss.org

Changes:

Initial Announcement on mloss.org.


Logo Local high order regularization 1.0

by kkim - March 2, 2016, 13:46:17 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 7498 views, 1649 downloads, 0 subscriptions

Rating Empty StarEmpty StarEmpty StarEmpty StarEmpty Star
(based on 1 vote)

About: Local high-order regularization for semi-supervised learning

Changes:

Initial Announcement on mloss.org.


Logo Sparse PCA 2.0

by tbuehler - December 31, 2015, 16:24:42 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 19723 views, 3846 downloads, 0 subscriptions

About: A Matlab implementation of Sparse PCA using the inverse power method for nonlinear eigenproblems.

Changes:
  • Added deflation scheme to compute multiple principal components
  • Several internal runtime and memory optimizations
  • API change: sparsePCA.m is now used to compute multiple components; use computeTradeOffCurve.m to reproduce the examples in the NIPS paper

Logo A Pattern Recognizer In Lua with ANNs v0.4.1

by pakozm - December 3, 2015, 15:01:36 CET [ Project Homepage BibTeX Download ] 27260 views, 6711 downloads, 0 subscriptions

About: APRIL-ANN toolkit (A Pattern Recognizer In Lua with Artificial Neural Networks). This toolkit incorporates ANN algorithms (as dropout, stacked denoising auto-encoders, convolutional neural networks), with other pattern recognition methods as hidden makov models (HMMs) among others.

Changes:
  • Updated home repository link to follow april-org github organization.
  • Improved serialize/deserialize functions, reimplemented all the serialization procedure.
  • Added exceptions support to LuaPkg and APRIL-ANN, allowing to capture C++ errors into Lua code.
  • Added set class.
  • Added series class.
  • Added data_frame class, similar to Python Pandas DataFrame.
  • Serialization and deserilization have been updated with more robust and reusable API, implemented in util.serialize() and util.deserialize() functions.
  • Added matrix.ext.broadcast utility (similar to broadcast in numpy).
  • Added ProbablisitcMatrixANNComponent, which allow to implement probabilistic mixtures of posteriors and/or likelihoods.
  • Added batch normalization ANN component.
  • Allowing matrix.join to add new axis.
  • Added methods prod(), cumsum() and cumprod() at matrix classes.
  • Added methods count_eq() and count_neq() at matrix classes.
  • Serializable objects API have been augmented with methods ctor_name() and
    ctor_params() in Lua, refered to luaCtorName() and luaCtorParams() in C++.
  • Added cast.to to dynamic cast C++ objects pushed into Lua, allowing to convert base class objects into any of its derived classes.
  • Added matrix.sparse as valid values for targets in ann.loss.mse and
    ann.loss.cross_entropy.
  • Changed matrix metamethods __index and __newindex, allowing to use
    matrix objects with standard Lua operator[].
  • Added matrix.masked_fill and matrix.masked_copy matrix.
  • Added matrix.indexed_fill and matrix.indexed_copy matrix.
  • Added ann.components.probabilistic_matrix, and its corresponding specializations ann.components.left_probabilistic_matrix and
    ann.components.right_probabilistic_matrix.
  • Added operator[] in the right side of matrix operations.
  • Added ann.components.transpose.
  • Added max_gradients_norm in traianble.supervised_trainer, to avoid gradients exploding.
  • Added ann.components.actf.sparse_logistic a logistic activation function with sparsity penalty.
  • Simplified math.add, math.sub, ... and other math extensions for reductions, their original behavior can be emulated by using bind function.
  • Added bind function to freeze any positional argument of any Lua function.
  • Function stats.boot uses multiple_unpack to allow a table of sizes and the generation of multiple index matrices.
  • Added multiple_unpack Lua function.
  • Added __tostring metamethod to numeric memory blocks in Lua.
  • Added dataset.token.sparse_matrix, a dataset which allow to traverse by rows a sparse matrix instance.
  • Added matrix.sparse.builders.dok, a builder which uses the Dictionary-of-Keys format to construct a sparse matrix from scratch.
  • Added method data to numeric matrix classes.
  • Added methods values, indices, first_index to sparse matrix class.
  • Fixed bugs when reading bad formed CSV files.
  • Fixed bugs at statistical distributions.
  • FloatRGB bug solved on equal (+=, -=, ...) operators. This bug affected ImageRGB operations such as resize.
  • Solved problems when chaining methods in Lua, some objects end to be garbage collected.
  • Improved support of strings in auto-completion (rlcompleter package).
  • Solved bug at SparseMatrix<T> when reading it from a file.
  • Solved bug in Image<T>::rotate90_cw methods.
  • Solved bug in SparseMatrix::toDense() method.

C/C++

  • Better LuaTable accessors, using [] operator.
  • Implementation of matrix __index, __newindex and __call metamethods in C++.
  • Implementation of matProd(), matCumSum() and matCumProd() functions.
  • Implementation of matCountEq() and matCountNeq() functions for
    Matrix<T>.
  • Updated matrix_ext_operations.h to change API of matrix operations. All functions have been overloaded to accept an in-place operation and another version which receives a destination matrix.
  • Adding iterators to language models.
  • Added MatrixScalarMap2 which receives as input2 a SparaseMatrix instance. This functions needs to be generalized to work with CPU and CUDA.
  • The method SparseMatrix<T>::fromDenseMatrix() uses a DOKBuilder object to build the sparse matrix.
  • The conversion of a Matrix<T> into a SparseMatrix<T> has been changed from a constructor overload to the static method
    SparseMatrix<T>::fromDenseMatrix().
  • Added support for IPyLua.
  • Optimized matrix access for confusion matrix.
  • Minor changes in class.lua.
  • Improved binding to avoid multiple object copies when pushing C++ objects.
  • Added Git commit hash and compilation time.

Logo A Library for Online Streaming Feature Selection 1.0

by ykui713 - November 25, 2015, 13:23:01 CET [ BibTeX Download ] 5135 views, 2211 downloads, 0 subscriptions

About: LOFS is a software toolbox for online streaming feature selection

Changes:

Initial Announcement on mloss.org.


Logo Cognitive Foundry 3.4.2

by Baz - October 30, 2015, 06:53:03 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 74988 views, 15079 downloads, 0 subscriptions

About: The Cognitive Foundry is a modular Java software library of machine learning components and algorithms designed for research and applications.

Changes:
  • General:
    • Upgraded MTJ to 1.0.3.
  • Common:
    • Added package for hash function computation including Eva, FNV-1a, MD5, Murmur2, Prime, SHA1, SHA2
    • Added callback-based forEach implementations to Vector and InfiniteVector, which can be faster for iterating through some vector types.
    • Optimized DenseVector by removing a layer of indirection.
    • Added method to compute set of percentiles in UnivariateStatisticsUtil and fixed issue with percentile interpolation.
    • Added utility class for enumerating combinations.
    • Adjusted ScalarMap implementation hierarchy.
    • Added method for copying a map to VectorFactory and moved createVectorCapacity up from SparseVectorFactory.
    • Added method for creating square identity matrix to MatrixFactory.
    • Added Random implementation that uses a cached set of values.
  • Learning:
    • Implemented feature hashing.
    • Added factory for random forests.
    • Implemented uniform distribution over integer values.
    • Added Chi-squared similarity.
    • Added KL divergence.
    • Added general conditional probability distribution.
    • Added interfaces for Regression, UnivariateRegression, and MultivariateRegression.
    • Fixed null pointer exception that can happen in K-means with an empty cluster.
    • Fixed name of maxClusters property on AgglomerativeClusterer (was called maxMinDistance).
  • Text:
    • Improvements to LDA Gibbs sampler.

Logo Chalearn gesture challenge code by jun wan 2.0

by joewan - September 29, 2015, 08:50:22 CET [ BibTeX BibTeX for corresponding Paper Download ] 17387 views, 3762 downloads, 0 subscriptions

About: This code is provided by Jun Wan. It is used in the Chalearn one-shot learning gesture challenge (round 2). This code includes: bag of features, 3D MoSIFT-based features (i.e. 3D MoSIFT, 3D EMoSIFT and 3D SMoSIFT), and the MFSK feature.

Changes:

Initial Announcement on mloss.org.


Showing Items 1-20 of 72 on page 1 of 4: 1 2 3 4 Next