See the version list below for details.
Install-Package DiffSharp -Version 0.7.7
dotnet add package DiffSharp --version 0.7.7
<PackageReference Include="DiffSharp" Version="0.7.7" />
paket add DiffSharp --version 0.7.7
#r "nuget: DiffSharp, 0.7.7"
// Install DiffSharp as a Cake Addin #addin nuget:?package=DiffSharp&version=0.7.7 // Install DiffSharp as a Cake Tool #tool nuget:?package=DiffSharp&version=0.7.7
DiffSharp is an automatic differentiation (AD) library.
AD allows exact and efficient calculation of derivatives, by systematically invoking the chain rule of calculus at the elementary operator level during program execution. AD is different from numerical differentiation, which is prone to truncation and round-off errors, and symbolic differentiation, which is affected by expression swell and cannot fully handle algorithmic control flow.
Using the DiffSharp library, derivative calculations (gradients, Hessians, Jacobians, directional derivatives, and matrix-free Hessian- and Jacobian-vector products) can be incorporated with minimal change into existing algorithms. Diffsharp supports nested forward and reverse AD up to any level, meaning that you can compute exact higher-order derivatives or differentiate functions that are internally making use of differentiation. Please see the API Overview page for a list of available operations.
The library is under active development by Atılım Güneş Baydin and Barak A. Pearlmutter mainly for research applications in machine learning, as part of their work at the Brain and Computation Lab, Hamilton Institute, National University of Ireland Maynooth.
DiffSharp is implemented in the F# language and can be used from C# and the other languages running on Mono or the .Net Framework, targeting the 64 bit platform. It is tested on Linux and Windows. We are working on interfaces/ports to other languages.
|.NET Framework||net46 net461 net462 net463 net47 net471 net472 net48|
- FSharp.Quotations.Evaluator (>= 1.0.6)
NuGet packages (1)
Showing the top 1 NuGet packages that depend on DiffSharp:
Hype is a proof-of-concept deep learning library, where you can perform optimization on compositional machine learning systems of many components, even when such components themselves internally perform optimization. This is enabled by nested automatic differentiation (AD) giving you access to the automatic exact derivative of any floating-point value in your code with respect to any other. Underlying computations are run by a BLAS/LAPACK backend (OpenBLAS by default).
This package is not used by any popular GitHub repositories.
for the latest release notes.