Reformat code

This commit is contained in:
Alexander Nozik 2024-03-27 09:11:12 +03:00
parent ecb5d28110
commit 214467d21c
200 changed files with 1023 additions and 591 deletions

View File

@ -77,7 +77,8 @@
- Major refactor of tensors (only minor API changes) - Major refactor of tensors (only minor API changes)
- Kotlin 1.8.20 - Kotlin 1.8.20
- `LazyStructure` `deffered` -> `async` to comply with coroutines code style - `LazyStructure` `deffered` -> `async` to comply with coroutines code style
- Default `dot` operation in tensor algebra no longer support broadcasting. Instead `matmul` operation is added to `DoubleTensorAlgebra`. - Default `dot` operation in tensor algebra no longer support broadcasting. Instead `matmul` operation is added
to `DoubleTensorAlgebra`.
- Multik went MPP - Multik went MPP
### Removed ### Removed
@ -236,9 +237,11 @@
- MST to JVM bytecode translator (https://github.com/mipt-npm/kmath/pull/94) - MST to JVM bytecode translator (https://github.com/mipt-npm/kmath/pull/94)
- FloatBuffer (specialized MutableBuffer over FloatArray) - FloatBuffer (specialized MutableBuffer over FloatArray)
- FlaggedBuffer to associate primitive numbers buffer with flags (to mark values infinite or missing, etc.) - FlaggedBuffer to associate primitive numbers buffer with flags (to mark values infinite or missing, etc.)
- Specialized builder functions for all primitive buffers like `IntBuffer(25) { it + 1 }` (https://github.com/mipt-npm/kmath/pull/125) - Specialized builder functions for all primitive buffers
like `IntBuffer(25) { it + 1 }` (https://github.com/mipt-npm/kmath/pull/125)
- Interface `NumericAlgebra` where `number` operation is available to convert numbers to algebraic elements - Interface `NumericAlgebra` where `number` operation is available to convert numbers to algebraic elements
- Inverse trigonometric functions support in ExtendedField (`asin`, `acos`, `atan`) (https://github.com/mipt-npm/kmath/pull/114) - Inverse trigonometric functions support in
ExtendedField (`asin`, `acos`, `atan`) (https://github.com/mipt-npm/kmath/pull/114)
- New space extensions: `average` and `averageWith` - New space extensions: `average` and `averageWith`
- Local coding conventions - Local coding conventions
- Geometric Domains API in `kmath-core` - Geometric Domains API in `kmath-core`
@ -251,7 +254,8 @@
- `readAsMemory` now has `throws IOException` in JVM signature. - `readAsMemory` now has `throws IOException` in JVM signature.
- Several functions taking functional types were made `inline`. - Several functions taking functional types were made `inline`.
- Several functions taking functional types now have `callsInPlace` contracts. - Several functions taking functional types now have `callsInPlace` contracts.
- BigInteger and BigDecimal algebra: JBigDecimalField has companion object with default math context; minor optimizations - BigInteger and BigDecimal algebra: JBigDecimalField has companion object with default math context; minor
optimizations
- `power(T, Int)` extension function has preconditions and supports `Field<T>` - `power(T, Int)` extension function has preconditions and supports `Field<T>`
- Memory objects have more preconditions (overflow checking) - Memory objects have more preconditions (overflow checking)
- `tg` function is renamed to `tan` (https://github.com/mipt-npm/kmath/pull/114) - `tg` function is renamed to `tan` (https://github.com/mipt-npm/kmath/pull/114)

127
README.md
View File

@ -25,7 +25,8 @@ experience could be achieved with [kmath-for-real](/kmath-for-real) extension mo
# Goal # Goal
* Provide a flexible and powerful API to work with mathematics abstractions in Kotlin-multiplatform (JVM, JS, Native and Wasm). * Provide a flexible and powerful API to work with mathematics abstractions in Kotlin-multiplatform (JVM, JS, Native and
Wasm).
* Provide basic multiplatform implementations for those abstractions (without significant performance optimization). * Provide basic multiplatform implementations for those abstractions (without significant performance optimization).
* Provide bindings and wrappers with those abstractions for popular optimized platform libraries. * Provide bindings and wrappers with those abstractions for popular optimized platform libraries.
@ -55,150 +56,181 @@ module definitions below. The module stability could have the following levels:
## Modules ## Modules
### [attributes-kt](attributes-kt) ### [attributes-kt](attributes-kt)
> An API and basic implementation for arranging objects in a continuous memory block. > An API and basic implementation for arranging objects in a continuous memory block.
> >
> **Maturity**: DEVELOPMENT > **Maturity**: DEVELOPMENT
### [benchmarks](benchmarks) ### [benchmarks](benchmarks)
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
### [examples](examples) ### [examples](examples)
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
### [kmath-ast](kmath-ast) ### [kmath-ast](kmath-ast)
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
> >
> **Features:** > **Features:**
> - [expression-language](kmath-ast/src/commonMain/kotlin/space/kscience/kmath/ast/parser.kt) : Expression language and its parser > - [expression-language](kmath-ast/src/commonMain/kotlin/space/kscience/kmath/ast/parser.kt) : Expression language and
> - [mst-jvm-codegen](kmath-ast/src/jvmMain/kotlin/space/kscience/kmath/asm/asm.kt) : Dynamic MST to JVM bytecode compiler its parser
> - [mst-jvm-codegen](kmath-ast/src/jvmMain/kotlin/space/kscience/kmath/asm/asm.kt) : Dynamic MST to JVM bytecode
compiler
> - [mst-js-codegen](kmath-ast/src/jsMain/kotlin/space/kscience/kmath/estree/estree.kt) : Dynamic MST to JS compiler > - [mst-js-codegen](kmath-ast/src/jsMain/kotlin/space/kscience/kmath/estree/estree.kt) : Dynamic MST to JS compiler
> - [rendering](kmath-ast/src/commonMain/kotlin/space/kscience/kmath/ast/rendering/MathRenderer.kt) : Extendable MST rendering > - [rendering](kmath-ast/src/commonMain/kotlin/space/kscience/kmath/ast/rendering/MathRenderer.kt) : Extendable MST
rendering
### [kmath-commons](kmath-commons) ### [kmath-commons](kmath-commons)
> Commons math binding for kmath > Commons math binding for kmath
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
### [kmath-complex](kmath-complex) ### [kmath-complex](kmath-complex)
> Complex numbers and quaternions. > Complex numbers and quaternions.
> >
> **Maturity**: PROTOTYPE > **Maturity**: PROTOTYPE
> >
> **Features:** > **Features:**
> - [complex](kmath-complex/src/commonMain/kotlin/space/kscience/kmath/complex/Complex.kt) : Complex numbers operations > - [complex](kmath-complex/src/commonMain/kotlin/space/kscience/kmath/complex/Complex.kt) : Complex numbers operations
> - [quaternion](kmath-complex/src/commonMain/kotlin/space/kscience/kmath/complex/Quaternion.kt) : Quaternions and their composition > - [quaternion](kmath-complex/src/commonMain/kotlin/space/kscience/kmath/complex/Quaternion.kt) : Quaternions and their
composition
### [kmath-core](kmath-core) ### [kmath-core](kmath-core)
> Core classes, algebra definitions, basic linear algebra > Core classes, algebra definitions, basic linear algebra
> >
> **Maturity**: DEVELOPMENT > **Maturity**: DEVELOPMENT
> >
> **Features:** > **Features:**
> - [algebras](kmath-core/src/commonMain/kotlin/space/kscience/kmath/operations/Algebra.kt) : Algebraic structures like rings, spaces and fields. > - [algebras](kmath-core/src/commonMain/kotlin/space/kscience/kmath/operations/Algebra.kt) : Algebraic structures like
> - [nd](kmath-core/src/commonMain/kotlin/space/kscience/kmath/structures/StructureND.kt) : Many-dimensional structures and operations on them. rings, spaces and fields.
> - [linear](kmath-core/src/commonMain/kotlin/space/kscience/kmath/operations/Algebra.kt) : Basic linear algebra operations (sums, products, etc.), backed by the `Space` API. Advanced linear algebra operations like matrix inversion and LU decomposition. > - [nd](kmath-core/src/commonMain/kotlin/space/kscience/kmath/structures/StructureND.kt) : Many-dimensional structures
and operations on them.
> - [linear](kmath-core/src/commonMain/kotlin/space/kscience/kmath/operations/Algebra.kt) : Basic linear algebra
operations (sums, products, etc.), backed by the `Space` API. Advanced linear algebra operations like matrix
inversion and LU decomposition.
> - [buffers](kmath-core/src/commonMain/kotlin/space/kscience/kmath/structures/Buffers.kt) : One-dimensional structure > - [buffers](kmath-core/src/commonMain/kotlin/space/kscience/kmath/structures/Buffers.kt) : One-dimensional structure
> - [expressions](kmath-core/src/commonMain/kotlin/space/kscience/kmath/expressions) : By writing a single mathematical expression once, users will be able to apply different types of > - [expressions](kmath-core/src/commonMain/kotlin/space/kscience/kmath/expressions) : By writing a single mathematical
expression once, users will be able to apply different types of
objects to the expression by providing a context. Expressions can be used for a wide variety of purposes from high objects to the expression by providing a context. Expressions can be used for a wide variety of purposes from high
performance calculations to code generation. performance calculations to code generation.
> - [domains](kmath-core/src/commonMain/kotlin/space/kscience/kmath/domains) : Domains > - [domains](kmath-core/src/commonMain/kotlin/space/kscience/kmath/domains) : Domains
> - [autodiff](kmath-core/src/commonMain/kotlin/space/kscience/kmath/expressions/SimpleAutoDiff.kt) : Automatic differentiation > - [autodiff](kmath-core/src/commonMain/kotlin/space/kscience/kmath/expressions/SimpleAutoDiff.kt) : Automatic
differentiation
### [kmath-coroutines](kmath-coroutines) ### [kmath-coroutines](kmath-coroutines)
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
### [kmath-dimensions](kmath-dimensions) ### [kmath-dimensions](kmath-dimensions)
> A proof of concept module for adding type-safe dimensions to structures > A proof of concept module for adding type-safe dimensions to structures
> >
> **Maturity**: PROTOTYPE > **Maturity**: PROTOTYPE
### [kmath-ejml](kmath-ejml) ### [kmath-ejml](kmath-ejml)
> >
> **Maturity**: PROTOTYPE > **Maturity**: PROTOTYPE
> >
> **Features:** > **Features:**
> - [ejml-vector](kmath-ejml/src/main/kotlin/space/kscience/kmath/ejml/EjmlVector.kt) : Point implementations. > - [ejml-vector](kmath-ejml/src/main/kotlin/space/kscience/kmath/ejml/EjmlVector.kt) : Point implementations.
> - [ejml-matrix](kmath-ejml/src/main/kotlin/space/kscience/kmath/ejml/EjmlMatrix.kt) : Matrix implementation. > - [ejml-matrix](kmath-ejml/src/main/kotlin/space/kscience/kmath/ejml/EjmlMatrix.kt) : Matrix implementation.
> - [ejml-linear-space](kmath-ejml/src/main/kotlin/space/kscience/kmath/ejml/EjmlLinearSpace.kt) : LinearSpace implementations. > - [ejml-linear-space](kmath-ejml/src/main/kotlin/space/kscience/kmath/ejml/EjmlLinearSpace.kt) : LinearSpace
implementations.
### [kmath-for-real](kmath-for-real) ### [kmath-for-real](kmath-for-real)
> Extension module that should be used to achieve numpy-like behavior. > Extension module that should be used to achieve numpy-like behavior.
All operations are specialized to work with `Double` numbers without declaring algebraic contexts. > All operations are specialized to work with `Double` numbers without declaring algebraic contexts.
One can still use generic algebras though. > One can still use generic algebras though.
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
> >
> **Features:** > **Features:**
> - [DoubleVector](kmath-for-real/src/commonMain/kotlin/space/kscience/kmath/real/DoubleVector.kt) : Numpy-like operations for Buffers/Points > - [DoubleVector](kmath-for-real/src/commonMain/kotlin/space/kscience/kmath/real/DoubleVector.kt) : Numpy-like
> - [DoubleMatrix](kmath-for-real/src/commonMain/kotlin/space/kscience/kmath/real/DoubleMatrix.kt) : Numpy-like operations for 2d real structures operations for Buffers/Points
> - [DoubleMatrix](kmath-for-real/src/commonMain/kotlin/space/kscience/kmath/real/DoubleMatrix.kt) : Numpy-like
operations for 2d real structures
> - [grids](kmath-for-real/src/commonMain/kotlin/space/kscience/kmath/structures/grids.kt) : Uniform grid generators > - [grids](kmath-for-real/src/commonMain/kotlin/space/kscience/kmath/structures/grids.kt) : Uniform grid generators
### [kmath-functions](kmath-functions) ### [kmath-functions](kmath-functions)
> Functions, integration and interpolation > Functions, integration and interpolation
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
> >
> **Features:** > **Features:**
> - [piecewise](kmath-functions/src/commonMain/kotlin/space/kscience/kmath/functions/Piecewise.kt) : Piecewise functions. > - [piecewise](kmath-functions/src/commonMain/kotlin/space/kscience/kmath/functions/Piecewise.kt) : Piecewise
> - [polynomials](kmath-functions/src/commonMain/kotlin/space/kscience/kmath/functions/Polynomial.kt) : Polynomial functions. functions.
> - [linear interpolation](kmath-functions/src/commonMain/kotlin/space/kscience/kmath/interpolation/LinearInterpolator.kt) : Linear XY interpolator. > - [polynomials](kmath-functions/src/commonMain/kotlin/space/kscience/kmath/functions/Polynomial.kt) : Polynomial
> - [spline interpolation](kmath-functions/src/commonMain/kotlin/space/kscience/kmath/interpolation/SplineInterpolator.kt) : Cubic spline XY interpolator. functions.
> - [linear interpolation](kmath-functions/src/commonMain/kotlin/space/kscience/kmath/interpolation/LinearInterpolator.kt) :
Linear XY interpolator.
> - [spline interpolation](kmath-functions/src/commonMain/kotlin/space/kscience/kmath/interpolation/SplineInterpolator.kt) :
Cubic spline XY interpolator.
> - [integration](kmath-functions/#) : Univariate and multivariate quadratures > - [integration](kmath-functions/#) : Univariate and multivariate quadratures
### [kmath-geometry](kmath-geometry) ### [kmath-geometry](kmath-geometry)
> >
> **Maturity**: PROTOTYPE > **Maturity**: PROTOTYPE
### [kmath-histograms](kmath-histograms) ### [kmath-histograms](kmath-histograms)
> >
> **Maturity**: PROTOTYPE > **Maturity**: PROTOTYPE
### [kmath-jafama](kmath-jafama) ### [kmath-jafama](kmath-jafama)
> Jafama integration module > Jafama integration module
> >
> **Maturity**: DEPRECATED > **Maturity**: DEPRECATED
> >
> **Features:** > **Features:**
> - [jafama-double](kmath-jafama/src/main/kotlin/space/kscience/kmath/jafama/) : Double ExtendedField implementations based on Jafama > - [jafama-double](kmath-jafama/src/main/kotlin/space/kscience/kmath/jafama/) : Double ExtendedField implementations
based on Jafama
### [kmath-jupyter](kmath-jupyter) ### [kmath-jupyter](kmath-jupyter)
> >
> **Maturity**: PROTOTYPE > **Maturity**: PROTOTYPE
### [kmath-kotlingrad](kmath-kotlingrad) ### [kmath-kotlingrad](kmath-kotlingrad)
> Kotlin∇ integration module > Kotlin∇ integration module
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
> >
> **Features:** > **Features:**
> - [differentiable-mst-expression](kmath-kotlingrad/src/main/kotlin/space/kscience/kmath/kotlingrad/KotlingradExpression.kt) : MST based DifferentiableExpression. > - [differentiable-mst-expression](kmath-kotlingrad/src/main/kotlin/space/kscience/kmath/kotlingrad/KotlingradExpression.kt) :
> - [scalars-adapters](kmath-kotlingrad/src/main/kotlin/space/kscience/kmath/kotlingrad/scalarsAdapters.kt) : Conversions between Kotlin∇'s SFun and MST MST based DifferentiableExpression.
> - [scalars-adapters](kmath-kotlingrad/src/main/kotlin/space/kscience/kmath/kotlingrad/scalarsAdapters.kt) :
Conversions between Kotlin∇'s SFun and MST
### [kmath-memory](kmath-memory) ### [kmath-memory](kmath-memory)
> An API and basic implementation for arranging objects in a continuous memory block. > An API and basic implementation for arranging objects in a continuous memory block.
> >
> **Maturity**: DEVELOPMENT > **Maturity**: DEVELOPMENT
### [kmath-multik](kmath-multik) ### [kmath-multik](kmath-multik)
> JetBrains Multik connector > JetBrains Multik connector
> >
> **Maturity**: PROTOTYPE > **Maturity**: PROTOTYPE
### [kmath-nd4j](kmath-nd4j) ### [kmath-nd4j](kmath-nd4j)
> ND4J NDStructure implementation and according NDAlgebra classes > ND4J NDStructure implementation and according NDAlgebra classes
> >
> **Maturity**: DEPRECATED > **Maturity**: DEPRECATED
@ -208,45 +240,52 @@ One can still use generic algebras though.
> - [nd4jarrayrings](kmath-nd4j/#) : Rings over Nd4jArrayStructure of Int and Long > - [nd4jarrayrings](kmath-nd4j/#) : Rings over Nd4jArrayStructure of Int and Long
> - [nd4jarrayfields](kmath-nd4j/#) : Fields over Nd4jArrayStructure of Float and Double > - [nd4jarrayfields](kmath-nd4j/#) : Fields over Nd4jArrayStructure of Float and Double
### [kmath-optimization](kmath-optimization) ### [kmath-optimization](kmath-optimization)
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
### [kmath-stat](kmath-stat) ### [kmath-stat](kmath-stat)
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
### [kmath-symja](kmath-symja) ### [kmath-symja](kmath-symja)
> Symja integration module > Symja integration module
> >
> **Maturity**: PROTOTYPE > **Maturity**: PROTOTYPE
### [kmath-tensorflow](kmath-tensorflow) ### [kmath-tensorflow](kmath-tensorflow)
> Google tensorflow connector > Google tensorflow connector
> >
> **Maturity**: PROTOTYPE > **Maturity**: PROTOTYPE
### [kmath-tensors](kmath-tensors) ### [kmath-tensors](kmath-tensors)
> >
> **Maturity**: PROTOTYPE > **Maturity**: PROTOTYPE
> >
> **Features:** > **Features:**
> - [tensor algebra](kmath-tensors/src/commonMain/kotlin/space/kscience/kmath/tensors/api/TensorAlgebra.kt) : Basic linear algebra operations on tensors (plus, dot, etc.) > - [tensor algebra](kmath-tensors/src/commonMain/kotlin/space/kscience/kmath/tensors/api/TensorAlgebra.kt) : Basic
> - [tensor algebra with broadcasting](kmath-tensors/src/commonMain/kotlin/space/kscience/kmath/tensors/core/BroadcastDoubleTensorAlgebra.kt) : Basic linear algebra operations implemented with broadcasting. linear algebra operations on tensors (plus, dot, etc.)
> - [linear algebra operations](kmath-tensors/src/commonMain/kotlin/space/kscience/kmath/tensors/api/LinearOpsTensorAlgebra.kt) : Advanced linear algebra operations like LU decomposition, SVD, etc. > - [tensor algebra with broadcasting](kmath-tensors/src/commonMain/kotlin/space/kscience/kmath/tensors/core/BroadcastDoubleTensorAlgebra.kt) :
Basic linear algebra operations implemented with broadcasting.
> - [linear algebra operations](kmath-tensors/src/commonMain/kotlin/space/kscience/kmath/tensors/api/LinearOpsTensorAlgebra.kt) :
Advanced linear algebra operations like LU decomposition, SVD, etc.
### [kmath-viktor](kmath-viktor) ### [kmath-viktor](kmath-viktor)
> Binding for https://github.com/JetBrains-Research/viktor > Binding for https://github.com/JetBrains-Research/viktor
> >
> **Maturity**: DEVELOPMENT > **Maturity**: DEVELOPMENT
### [test-utils](test-utils) ### [test-utils](test-utils)
> >
> **Maturity**: EXPERIMENTAL > **Maturity**: EXPERIMENTAL
## Multi-platform support ## Multi-platform support
KMath is developed as a multi-platform library, which means that most of the interfaces are declared in the KMath is developed as a multi-platform library, which means that most of the interfaces are declared in the
@ -257,16 +296,19 @@ feedback are also welcome.
## Performance ## Performance
Calculation of performance is one of the major goals of KMath in the future, but in some cases it is impossible to achieve both Calculation of performance is one of the major goals of KMath in the future, but in some cases it is impossible to
achieve both
performance and flexibility. performance and flexibility.
We expect to focus on creating a convenient universal API first and then work on increasing performance for specific We expect to focus on creating a convenient universal API first and then work on increasing performance for specific
cases. We expect the worst KMath benchmarks will perform better than native Python, but worse than optimized cases. We expect the worst KMath benchmarks will perform better than native Python, but worse than optimized
native/SciPy (mostly due to boxing operations on primitive numbers). The best performance of optimized parts could be better than SciPy. native/SciPy (mostly due to boxing operations on primitive numbers). The best performance of optimized parts could be
better than SciPy.
## Requirements ## Requirements
KMath currently relies on JDK 11 for compilation and execution of Kotlin-JVM part. We recommend using GraalVM-CE or Oracle GraalVM for execution to get better performance. KMath currently relies on JDK 11 for compilation and execution of Kotlin-JVM part. We recommend using GraalVM-CE or
Oracle GraalVM for execution to get better performance.
### Repositories ### Repositories
@ -289,4 +331,7 @@ dependencies {
## Contributing ## Contributing
The project requires a lot of additional work. The most important thing we need is feedback about what features are The project requires a lot of additional work. The most important thing we need is feedback about what features are
required the most. Feel free to create feature requests. We are also welcome to code contributions, especially in issues marked with [good first issue](hhttps://github.com/SciProgCentre/kmath/issues?q=is%3Aissue+is%3Aopen+label%3A%22good+first+issue%22) label. required the most. Feel free to create feature requests. We are also welcome to code contributions, especially in issues
marked
with [good first issue](hhttps://github.com/SciProgCentre/kmath/issues?q=is%3Aissue+is%3Aopen+label%3A%22good+first+issue%22)
label.

View File

@ -22,7 +22,8 @@ public abstract class PolymorphicAttribute<T>(public val type: SafeType<T>) : At
* Get a polymorphic attribute using attribute factory * Get a polymorphic attribute using attribute factory
*/ */
@UnstableAttributesAPI @UnstableAttributesAPI
public operator fun <T> Attributes.get(attributeKeyBuilder: () -> PolymorphicAttribute<T>): T? = get(attributeKeyBuilder()) public operator fun <T> Attributes.get(attributeKeyBuilder: () -> PolymorphicAttribute<T>): T? =
get(attributeKeyBuilder())
/** /**
* Set a polymorphic attribute using its factory * Set a polymorphic attribute using its factory

View File

@ -94,6 +94,7 @@ class ExpressionsInterpretersBenchmark {
} }
private val mst = node.toExpression(Float64Field) private val mst = node.toExpression(Float64Field)
@OptIn(UnstableKMathAPI::class) @OptIn(UnstableKMathAPI::class)
private val wasm = node.wasmCompileToExpression(Float64Field) private val wasm = node.wasmCompileToExpression(Float64Field)
private val estree = node.estreeCompileToExpression(Float64Field) private val estree = node.estreeCompileToExpression(Float64Field)

View File

@ -63,7 +63,8 @@ fun Project.addBenchmarkProperties() {
if (resDirectory == null || !(resDirectory.resolve("jvm.json")).exists()) { if (resDirectory == null || !(resDirectory.resolve("jvm.json")).exists()) {
"> **Can't find appropriate benchmark data. Try generating readme files after running benchmarks**." "> **Can't find appropriate benchmark data. Try generating readme files after running benchmarks**."
} else { } else {
val reports: List<JmhReport> = jsonMapper.readValue<List<JmhReport>>(resDirectory.resolve("jvm.json")) val reports: List<JmhReport> =
jsonMapper.readValue<List<JmhReport>>(resDirectory.resolve("jvm.json"))
buildString { buildString {
appendLine("<details>") appendLine("<details>")
@ -76,16 +77,20 @@ fun Project.addBenchmarkProperties() {
appendLine("* Run on ${first.vmName} (build ${first.vmVersion}) with Java process:") appendLine("* Run on ${first.vmName} (build ${first.vmVersion}) with Java process:")
appendLine() appendLine()
appendLine("```") appendLine("```")
appendLine("${first.jvm} ${ appendLine(
"${first.jvm} ${
first.jvmArgs.joinToString(" ") first.jvmArgs.joinToString(" ")
}") }"
)
appendLine("```") appendLine("```")
appendLine("* JMH ${first.jmhVersion} was used in `${first.mode}` mode with ${first.warmupIterations} warmup ${ appendLine(
"* JMH ${first.jmhVersion} was used in `${first.mode}` mode with ${first.warmupIterations} warmup ${
noun(first.warmupIterations, "iteration", "iterations") noun(first.warmupIterations, "iteration", "iterations")
} by ${first.warmupTime} and ${first.measurementIterations} measurement ${ } by ${first.warmupTime} and ${first.measurementIterations} measurement ${
noun(first.measurementIterations, "iteration", "iterations") noun(first.measurementIterations, "iteration", "iterations")
} by ${first.measurementTime}.") } by ${first.measurementTime}."
)
appendLine() appendLine()
appendLine("| Benchmark | Score |") appendLine("| Benchmark | Score |")

View File

@ -1,20 +1,27 @@
# Coding Conventions # Coding Conventions
Generally, KMath code follows general [Kotlin coding conventions](https://kotlinlang.org/docs/reference/coding-conventions.html), but with a number of small changes and clarifications. Generally, KMath code follows
general [Kotlin coding conventions](https://kotlinlang.org/docs/reference/coding-conventions.html), but with a number of
small changes and clarifications.
## Utility Class Naming ## Utility Class Naming
Filename should coincide with a name of one of the classes contained in the file or start with small letter and describe its contents. Filename should coincide with a name of one of the classes contained in the file or start with small letter and describe
its contents.
The code convention [here](https://kotlinlang.org/docs/reference/coding-conventions.html#source-file-names) says that file names should start with a capital letter even if file does not contain classes. Yet starting utility classes and aggregators with a small letter seems to be a good way to visually separate those files. The code convention [here](https://kotlinlang.org/docs/reference/coding-conventions.html#source-file-names) says that
file names should start with a capital letter even if file does not contain classes. Yet starting utility classes and
aggregators with a small letter seems to be a good way to visually separate those files.
This convention could be changed in future in a non-breaking way. This convention could be changed in future in a non-breaking way.
## Private Variable Naming ## Private Variable Naming
Private variables' names may start with underscore `_` for of the private mutable variable is shadowed by the public read-only value with the same meaning. Private variables' names may start with underscore `_` for of the private mutable variable is shadowed by the public
read-only value with the same meaning.
This rule does not permit underscores in names, but it is sometimes useful to "underscore" the fact that public and private versions draw up the same entity. It is allowed only for private variables. This rule does not permit underscores in names, but it is sometimes useful to "underscore" the fact that public and
private versions draw up the same entity. It is allowed only for private variables.
This convention could be changed in future in a non-breaking way. This convention could be changed in future in a non-breaking way.
@ -24,4 +31,5 @@ Use one-liners when they occupy single code window line both for functions and p
`val b: String get() = "fff"`. The same should be performed with multiline expressions when they could be `val b: String get() = "fff"`. The same should be performed with multiline expressions when they could be
cleanly separated. cleanly separated.
There is no universal consensus whenever use `fun a() = ...` or `fun a() { return ... }`. Yet from reader outlook one-lines seem to better show that the property or function is easily calculated. There is no universal consensus whenever use `fun a() = ...` or `fun a() { return ... }`. Yet from reader outlook
one-lines seem to better show that the property or function is easily calculated.

View File

@ -1,16 +1,19 @@
# Expressions # Expressions
Expressions is a feature, which allows constructing lazily or immediately calculated parametric mathematical expressions. Expressions is a feature, which allows constructing lazily or immediately calculated parametric mathematical
expressions.
The potential use-cases for it (so far) are following: The potential use-cases for it (so far) are following:
* lazy evaluation (in general simple lambda is better, but there are some border cases); * lazy evaluation (in general simple lambda is better, but there are some border cases);
* automatic differentiation in single-dimension and in multiple dimensions; * automatic differentiation in single-dimension and in multiple dimensions;
* generation of mathematical syntax trees with subsequent code generation for other languages; * generation of mathematical syntax trees with subsequent code generation for other languages;
* symbolic computations, especially differentiation (and some other actions with `kmath-symja` integration with Symja's `IExpr`&mdash;integration, simplification, and more); * symbolic computations, especially differentiation (and some other actions with `kmath-symja` integration with
Symja's `IExpr`&mdash;integration, simplification, and more);
* visualization with `kmath-jupyter`. * visualization with `kmath-jupyter`.
The workhorse of this API is `Expression` interface, which exposes single `operator fun invoke(arguments: Map<Symbol, T>): T` The workhorse of this API is `Expression` interface, which exposes
single `operator fun invoke(arguments: Map<Symbol, T>): T`
method. `ExpressionAlgebra` is used to generate expressions and introduce variables. method. `ExpressionAlgebra` is used to generate expressions and introduce variables.
Currently there are two implementations: Currently there are two implementations:

View File

@ -1,8 +1,12 @@
## Basic linear algebra layout ## Basic linear algebra layout
KMath support for linear algebra organized in a context-oriented way, which means that operations are in most cases declared in context classes, and are not the members of classes that store data. This allows more flexible approach to maintain multiple back-ends. The new operations added as extensions to contexts instead of being member functions of data structures. KMath support for linear algebra organized in a context-oriented way, which means that operations are in most cases
declared in context classes, and are not the members of classes that store data. This allows more flexible approach to
maintain multiple back-ends. The new operations added as extensions to contexts instead of being member functions of
data structures.
The main context for linear algebra over matrices and vectors is `LinearSpace`, which defines addition and dot products of matrices and vectors: The main context for linear algebra over matrices and vectors is `LinearSpace`, which defines addition and dot products
of matrices and vectors:
```kotlin ```kotlin
import space.kscience.kmath.linear.* import space.kscience.kmath.linear.*
@ -28,4 +32,5 @@ LinearSpace.Companion.real {
## Backends overview ## Backends overview
### EJML ### EJML
### Commons Math ### Commons Math

View File

@ -8,6 +8,7 @@ One of the most sought after features of mathematical libraries is the high-perf
structures. In `kmath` performance depends on which particular context was used for operation. structures. In `kmath` performance depends on which particular context was used for operation.
Let us consider following contexts: Let us consider following contexts:
```kotlin ```kotlin
// automatically build context most suited for given type. // automatically build context most suited for given type.
val autoField = NDField.auto(DoubleField, dim, dim) val autoField = NDField.auto(DoubleField, dim, dim)
@ -16,6 +17,7 @@ Let us consider following contexts:
//A generic boxing field. It should be used for objects, not primitives. //A generic boxing field. It should be used for objects, not primitives.
val genericField = NDField.buffered(DoubleField, dim, dim) val genericField = NDField.buffered(DoubleField, dim, dim)
``` ```
Now let us perform several tests and see, which implementation is best suited for each case: Now let us perform several tests and see, which implementation is best suited for each case:
## Test case ## Test case
@ -24,7 +26,9 @@ To test performance we will take 2d-structures with `dim = 1000` and add a struc
to it `n = 1000` times. to it `n = 1000` times.
## Specialized ## Specialized
The code to run this looks like: The code to run this looks like:
```kotlin ```kotlin
specializedField.run { specializedField.run {
var res: NDBuffer<Double> = one var res: NDBuffer<Double> = one
@ -33,13 +37,16 @@ The code to run this looks like:
} }
} }
``` ```
The performance of this code is the best of all tests since it inlines all operations and is specialized for operation The performance of this code is the best of all tests since it inlines all operations and is specialized for operation
with doubles. We will measure everything else relative to this one, so time for this test will be `1x` (real time with doubles. We will measure everything else relative to this one, so time for this test will be `1x` (real time
on my computer is about 4.5 seconds). The only problem with this approach is that it requires specifying type on my computer is about 4.5 seconds). The only problem with this approach is that it requires specifying type
from the beginning. Everyone does so anyway, so it is the recommended approach. from the beginning. Everyone does so anyway, so it is the recommended approach.
## Automatic ## Automatic
Let's do the same with automatic field inference: Let's do the same with automatic field inference:
```kotlin ```kotlin
autoField.run { autoField.run {
var res = one var res = one
@ -48,13 +55,16 @@ Let's do the same with automatic field inference:
} }
} }
``` ```
Ths speed of this operation is approximately the same as for specialized case since `NDField.auto` just Ths speed of this operation is approximately the same as for specialized case since `NDField.auto` just
returns the same `RealNDField` in this case. Of course, it is usually better to use specialized method to be sure. returns the same `RealNDField` in this case. Of course, it is usually better to use specialized method to be sure.
## Lazy ## Lazy
Lazy field does not produce a structure when asked, instead it generates an empty structure and fills it on-demand Lazy field does not produce a structure when asked, instead it generates an empty structure and fills it on-demand
using coroutines to parallelize computations. using coroutines to parallelize computations.
When one calls When one calls
```kotlin ```kotlin
lazyField.run { lazyField.run {
var res = one var res = one
@ -63,12 +73,14 @@ When one calls
} }
} }
``` ```
The result will be calculated almost immediately but the result will be empty. To get the full result The result will be calculated almost immediately but the result will be empty. To get the full result
structure one needs to call all its elements. In this case computation overhead will be huge. So this field never structure one needs to call all its elements. In this case computation overhead will be huge. So this field never
should be used if one expects to use the full result structure. Though if one wants only small fraction, it could should be used if one expects to use the full result structure. Though if one wants only small fraction, it could
save a lot of time. save a lot of time.
This field still could be used with reasonable performance if call code is changed: This field still could be used with reasonable performance if call code is changed:
```kotlin ```kotlin
lazyField.run { lazyField.run {
val res = one.map { val res = one.map {
@ -82,10 +94,13 @@ This field still could be used with reasonable performance if call code is chang
res.elements().forEach { it.second } res.elements().forEach { it.second }
} }
``` ```
In this case it completes in about `4x-5x` time due to boxing. In this case it completes in about `4x-5x` time due to boxing.
## Boxing ## Boxing
The boxing field produced by The boxing field produced by
```kotlin ```kotlin
genericField.run { genericField.run {
var res: NDBuffer<Double> = one var res: NDBuffer<Double> = one
@ -94,18 +109,22 @@ The boxing field produced by
} }
} }
``` ```
is the slowest one, because it requires boxing and unboxing the `double` on each operation. It takes about is the slowest one, because it requires boxing and unboxing the `double` on each operation. It takes about
`15x` time (**TODO: there seems to be a problem here, it should be slow, but not that slow**). This field should `15x` time (**TODO: there seems to be a problem here, it should be slow, but not that slow**). This field should
never be used for primitives. never be used for primitives.
## Element operation ## Element operation
Let us also check the speed for direct operations on elements: Let us also check the speed for direct operations on elements:
```kotlin ```kotlin
var res = genericField.one var res = genericField.one
repeat(n) { repeat(n) {
res += 1.0 res += 1.0
} }
``` ```
One would expect to be at least as slow as field operation, but in fact, this one takes only `2x` time to complete. One would expect to be at least as slow as field operation, but in fact, this one takes only `2x` time to complete.
It happens, because in this particular case it does not use actual `NDField` but instead calculated directly It happens, because in this particular case it does not use actual `NDField` but instead calculated directly
via extension function. via extension function.
@ -114,6 +133,7 @@ via extension function.
Usually it is bad idea to compare the direct numerical operation performance in different languages, but it hard to Usually it is bad idea to compare the direct numerical operation performance in different languages, but it hard to
work completely without frame of reference. In this case, simple numpy code: work completely without frame of reference. In this case, simple numpy code:
```python ```python
import numpy as np import numpy as np
@ -121,7 +141,9 @@ res = np.ones((1000,1000))
for i in range(1000): for i in range(1000):
res = res + 1.0 res = res + 1.0
``` ```
gives the completion time of about `1.1x`, which means that specialized kotlin code in fact is working faster (I think it is
gives the completion time of about `1.1x`, which means that specialized kotlin code in fact is working faster (I think
it is
because better memory management). Of course if one writes `res += 1.0`, the performance will be different, because better memory management). Of course if one writes `res += 1.0`, the performance will be different,
but it would be different case, because numpy overrides `+=` with in-place operations. In-place operations are but it would be different case, because numpy overrides `+=` with in-place operations. In-place operations are
available in `kmath` with `MutableNDStructure` but there is no field for it (one can still work with mapping available in `kmath` with `MutableNDStructure` but there is no field for it (one can still work with mapping

View File

@ -1,27 +1,54 @@
# Polynomials and Rational Functions # Polynomials and Rational Functions
KMath provides a way to work with uni- and multivariate polynomials and rational functions. It includes full support of arithmetic operations of integers, **constants** (elements of ring polynomials are build over), variables (for certain multivariate implementations), polynomials and rational functions encapsulated in so-called **polynomial space** and **rational function space** and some other utilities such as algebraic differentiation and substitution. KMath provides a way to work with uni- and multivariate polynomials and rational functions. It includes full support of
arithmetic operations of integers, **constants** (elements of ring polynomials are build over), variables (for certain
multivariate implementations), polynomials and rational functions encapsulated in so-called **polynomial space** and *
*rational function space** and some other utilities such as algebraic differentiation and substitution.
## Concrete realizations ## Concrete realizations
There are 3 approaches to represent polynomials: There are 3 approaches to represent polynomials:
1. For univariate polynomials one can represent and store polynomial as a list of coefficients for each power of the variable. I.e. polynomial $a_0 + \dots + a_n x^n $ can be represented as a finite sequence $(a_0; \dots; a_n)$. (Compare to sequential definition of polynomials.)
2. For multivariate polynomials one can represent and store polynomial as a matching (in programming it is called "map" or "dictionary", in math it is called [functional relation](https://en.wikipedia.org/wiki/Binary_relation#Special_types_of_binary_relations)) of each "**term signature**" (that describes what variables and in what powers appear in the term) with corresponding coefficient of the term. But there are 2 possible approaches of term signature representation:
1. One can number all the variables, so term signature can be represented as a sequence describing powers of the variables. I.e. signature of term $c \\; x_0^{d_0} \dots x_n^{d_n} $ (for natural or zero $d_i $) can be represented as a finite sequence $(d_0; \dots; d_n)$.
2. One can represent variables as objects ("**labels**"), so term signature can be also represented as a matching of each appeared variable with its power in the term. I.e. signature of term $c \\; x_0^{d_0} \dots x_n^{d_n} $ (for natural non-zero $d_i $) can be represented as a finite matching $(x_0 \to d_1; \dots; x_n \to d_n)$.
All that three approaches are implemented by "list", "numbered", and "labeled" versions of polynomials and polynomial spaces respectively. Whereas all rational functions are represented as fractions with corresponding polynomial numerator and denominator, and rational functions' spaces are implemented in the same way as usual field of rational numbers (or more precisely, as any field of fractions over integral domain) should be implemented. 1. For univariate polynomials one can represent and store polynomial as a list of coefficients for each power of the
variable. I.e. polynomial $a_0 + \dots + a_n x^n $ can be represented as a finite sequence $(a_0; \dots; a_n)$. (
Compare to sequential definition of polynomials.)
2. For multivariate polynomials one can represent and store polynomial as a matching (in programming it is called "map"
or "dictionary", in math it is
called [functional relation](https://en.wikipedia.org/wiki/Binary_relation#Special_types_of_binary_relations)) of
each "**term signature**" (that describes what variables and in what powers appear in the term) with corresponding
coefficient of the term. But there are 2 possible approaches of term signature representation:
1. One can number all the variables, so term signature can be represented as a sequence describing powers of the
variables. I.e. signature of term $c \\; x_0^{d_0} \dots x_n^{d_n} $ (for natural or zero $d_i $) can be
represented as a finite sequence $(d_0; \dots; d_n)$.
2. One can represent variables as objects ("**labels**"), so term signature can be also represented as a matching of
each appeared variable with its power in the term. I.e. signature of term $c \\; x_0^{d_0} \dots x_n^{d_n} $ (for
natural non-zero $d_i $) can be represented as a finite matching $(x_0 \to d_1; \dots; x_n \to d_n)$.
All that three approaches are implemented by "list", "numbered", and "labeled" versions of polynomials and polynomial
spaces respectively. Whereas all rational functions are represented as fractions with corresponding polynomial numerator
and denominator, and rational functions' spaces are implemented in the same way as usual field of rational numbers (or
more precisely, as any field of fractions over integral domain) should be implemented.
So here are a bit of details. Let `C` by type of constants. Then: So here are a bit of details. Let `C` by type of constants. Then:
1. `ListPolynomial`, `ListPolynomialSpace`, `ListRationalFunction` and `ListRationalFunctionSpace` implement the first scenario. `ListPolynomial` stores polynomial $a_0 + \dots + a_n x^n $ as a coefficients list `listOf(a_0, ..., a_n)` (of type `List<C>`).
They also have variation `ScalableListPolynomialSpace` that replaces former polynomials and implements `ScaleOperations`. 1. `ListPolynomial`, `ListPolynomialSpace`, `ListRationalFunction` and `ListRationalFunctionSpace` implement the first
2. `NumberedPolynomial`, `NumberedPolynomialSpace`, `NumberedRationalFunction` and `NumberedRationalFunctionSpace` implement second scenario. `NumberedPolynomial` stores polynomials as structures of type `Map<List<UInt>, C>`. Signatures are stored as `List<UInt>`. To prevent ambiguity signatures should not end with zeros. scenario. `ListPolynomial` stores polynomial $a_0 + \dots + a_n x^n $ as a coefficients
3. `LabeledPolynomial`, `LabeledPolynomialSpace`, `LabeledRationalFunction` and `LabeledRationalFunctionSpace` implement third scenario using common `Symbol` as variable type. `LabeledPolynomial` stores polynomials as structures of type `Map<Map<Symbol, UInt>, C>`. Signatures are stored as `Map<Symbol, UInt>`. To prevent ambiguity each signature should not map any variable to zero. list `listOf(a_0, ..., a_n)` (of type `List<C>`).
They also have variation `ScalableListPolynomialSpace` that replaces former polynomials and
implements `ScaleOperations`.
2. `NumberedPolynomial`, `NumberedPolynomialSpace`, `NumberedRationalFunction` and `NumberedRationalFunctionSpace`
implement second scenario. `NumberedPolynomial` stores polynomials as structures of type `Map<List<UInt>, C>`.
Signatures are stored as `List<UInt>`. To prevent ambiguity signatures should not end with zeros.
3. `LabeledPolynomial`, `LabeledPolynomialSpace`, `LabeledRationalFunction` and `LabeledRationalFunctionSpace` implement
third scenario using common `Symbol` as variable type. `LabeledPolynomial` stores polynomials as structures of
type `Map<Map<Symbol, UInt>, C>`. Signatures are stored as `Map<Symbol, UInt>`. To prevent ambiguity each signature
should not map any variable to zero.
### Example: `ListPolynomial` ### Example: `ListPolynomial`
For example, polynomial $2 - 3x + x^2 $ (with `Int` coefficients) is represented For example, polynomial $2 - 3x + x^2 $ (with `Int` coefficients) is represented
```kotlin ```kotlin
val polynomial: ListPolynomial<Int> = ListPolynomial(listOf(2, -3, 1)) val polynomial: ListPolynomial<Int> = ListPolynomial(listOf(2, -3, 1))
// or // or
@ -29,6 +56,7 @@ val polynomial: ListPolynomial<Int> = ListPolynomial(2, -3, 1)
``` ```
All algebraic operations can be used in corresponding space: All algebraic operations can be used in corresponding space:
```kotlin ```kotlin
val computationResult = Int.algebra.listPolynomialSpace { val computationResult = Int.algebra.listPolynomialSpace {
ListPolynomial(2, -3, 1) + ListPolynomial(0, 6) == ListPolynomial(2, 3, 1) ListPolynomial(2, -3, 1) + ListPolynomial(0, 6) == ListPolynomial(2, 3, 1)
@ -42,6 +70,7 @@ For more see [examples](../examples/src/main/kotlin/space/kscience/kmath/functio
### Example: `NumberedPolynomial` ### Example: `NumberedPolynomial`
For example, polynomial $3 + 5 x_1 - 7 x_0^2 x_2 $ (with `Int` coefficients) is represented For example, polynomial $3 + 5 x_1 - 7 x_0^2 x_2 $ (with `Int` coefficients) is represented
```kotlin ```kotlin
val polynomial: NumberedPolynomial<Int> = NumberedPolynomial( val polynomial: NumberedPolynomial<Int> = NumberedPolynomial(
mapOf( mapOf(
@ -59,6 +88,7 @@ val polynomial: NumberedPolynomial<Int> = NumberedPolynomial(
``` ```
All algebraic operations can be used in corresponding space: All algebraic operations can be used in corresponding space:
```kotlin ```kotlin
val computationResult = Int.algebra.numberedPolynomialSpace { val computationResult = Int.algebra.numberedPolynomialSpace {
NumberedPolynomial( NumberedPolynomial(
@ -84,6 +114,7 @@ For more see [examples](../examples/src/main/kotlin/space/kscience/kmath/functio
### Example: `LabeledPolynomial` ### Example: `LabeledPolynomial`
For example, polynomial $3 + 5 y - 7 x^2 z $ (with `Int` coefficients) is represented For example, polynomial $3 + 5 y - 7 x^2 z $ (with `Int` coefficients) is represented
```kotlin ```kotlin
val polynomial: LabeledPolynomial<Int> = LabeledPolynomial( val polynomial: LabeledPolynomial<Int> = LabeledPolynomial(
mapOf( mapOf(
@ -101,6 +132,7 @@ val polynomial: LabeledPolynomial<Int> = LabeledPolynomial(
``` ```
All algebraic operations can be used in corresponding space: All algebraic operations can be used in corresponding space:
```kotlin ```kotlin
val computationResult = Int.algebra.labeledPolynomialSpace { val computationResult = Int.algebra.labeledPolynomialSpace {
LabeledPolynomial( LabeledPolynomial(
@ -150,23 +182,42 @@ classDiagram
PolynomialSpaceOfFractions <|-- MultivariatePolynomialSpaceOfFractions PolynomialSpaceOfFractions <|-- MultivariatePolynomialSpaceOfFractions
``` ```
There are implemented `Polynomial` and `RationalFunction` interfaces as abstractions of polynomials and rational functions respectively (although, there is not a lot of logic in them) and `PolynomialSpace` and `RationalFunctionSpace` (that implement `Ring` interface) as abstractions of polynomials' and rational functions' spaces respectively. More precisely, that means they allow to declare common logic of interaction with such objects and spaces: There are implemented `Polynomial` and `RationalFunction` interfaces as abstractions of polynomials and rational
functions respectively (although, there is not a lot of logic in them) and `PolynomialSpace`
and `RationalFunctionSpace` (that implement `Ring` interface) as abstractions of polynomials' and rational functions'
spaces respectively. More precisely, that means they allow to declare common logic of interaction with such objects and
spaces:
- `Polynomial` does not provide any logic. It is marker interface. - `Polynomial` does not provide any logic. It is marker interface.
- `RationalFunction` provides numerator and denominator of rational function and destructuring declaration for them. - `RationalFunction` provides numerator and denominator of rational function and destructuring declaration for them.
- `PolynomialSpace` provides all possible arithmetic interactions of integers, constants (of type `C`), and polynomials (of type `P`) like addition, subtraction, multiplication, and some others and common properties like degree of polynomial. - `PolynomialSpace` provides all possible arithmetic interactions of integers, constants (of type `C`), and
- `RationalFunctionSpace` provides the same as `PolynomialSpace` but also for rational functions: all possible arithmetic interactions of integers, constants (of type `C`), polynomials (of type `P`), and rational functions (of type `R`) like addition, subtraction, multiplication, division (in some cases), and some others and common properties like degree of polynomial. polynomials (of type `P`) like addition, subtraction, multiplication, and some others and common properties like
degree of polynomial.
- `RationalFunctionSpace` provides the same as `PolynomialSpace` but also for rational functions: all possible
arithmetic interactions of integers, constants (of type `C`), polynomials (of type `P`), and rational functions (of
type `R`) like addition, subtraction, multiplication, division (in some cases), and some others and common properties
like degree of polynomial.
Then to add abstraction of similar behaviour with variables (in multivariate case) there are implemented `MultivariatePolynomialSpace` and `MultivariateRationalFunctionSpace`. They just include variables (of type `V`) in the interactions of the entities. Then to add abstraction of similar behaviour with variables (in multivariate case) there are
implemented `MultivariatePolynomialSpace` and `MultivariateRationalFunctionSpace`. They just include variables (of
type `V`) in the interactions of the entities.
Also, to remove boilerplates there were provided helping subinterfaces and abstract subclasses: Also, to remove boilerplates there were provided helping subinterfaces and abstract subclasses:
- `PolynomialSpaceOverRing` allows to replace implementation of interactions of integers and constants with implementations from provided ring over constants (of type `A: Ring<C>`).
- `PolynomialSpaceOverRing` allows to replace implementation of interactions of integers and constants with
implementations from provided ring over constants (of type `A: Ring<C>`).
- `RationalFunctionSpaceOverRing` &mdash; the same but for `RationalFunctionSpace`. - `RationalFunctionSpaceOverRing` &mdash; the same but for `RationalFunctionSpace`.
- `RationalFunctionSpaceOverPolynomialSpace` &mdash; the same but "the inheritance" includes interactions with polynomials from provided `PolynomialSpace`. - `RationalFunctionSpaceOverPolynomialSpace` &mdash; the same but "the inheritance" includes interactions with
- `PolynomialSpaceOfFractions` is actually abstract subclass of `RationalFunctionSpace` that implements all fractions boilerplates with provided (`protected`) constructor of rational functions by polynomial numerator and denominator. polynomials from provided `PolynomialSpace`.
- `MultivariateRationalFunctionSpaceOverMultivariatePolynomialSpace` and `MultivariatePolynomialSpaceOfFractions` &mdash; the same stories of operators inheritance and fractions boilerplates respectively but in multivariate case. - `PolynomialSpaceOfFractions` is actually abstract subclass of `RationalFunctionSpace` that implements all fractions
boilerplates with provided (`protected`) constructor of rational functions by polynomial numerator and denominator.
- `MultivariateRationalFunctionSpaceOverMultivariatePolynomialSpace` and `MultivariatePolynomialSpaceOfFractions`
&mdash; the same stories of operators inheritance and fractions boilerplates respectively but in multivariate case.
## Utilities ## Utilities
For all kinds of polynomials there are provided (implementation details depend on kind of polynomials) such common utilities as: For all kinds of polynomials there are provided (implementation details depend on kind of polynomials) such common
utilities as:
1. differentiation and anti-differentiation, 1. differentiation and anti-differentiation,
2. substitution, invocation and functional representation. 2. substitution, invocation and functional representation.

View File

@ -3,6 +3,7 @@
The Maven coordinates of this project are `${group}:${name}:${version}`. The Maven coordinates of this project are `${group}:${name}:${version}`.
**Gradle:** **Gradle:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -25,7 +25,8 @@ experience could be achieved with [kmath-for-real](/kmath-for-real) extension mo
# Goal # Goal
* Provide a flexible and powerful API to work with mathematics abstractions in Kotlin-multiplatform (JVM, JS, Native and Wasm). * Provide a flexible and powerful API to work with mathematics abstractions in Kotlin-multiplatform (JVM, JS, Native and
Wasm).
* Provide basic multiplatform implementations for those abstractions (without significant performance optimization). * Provide basic multiplatform implementations for those abstractions (without significant performance optimization).
* Provide bindings and wrappers with those abstractions for popular optimized platform libraries. * Provide bindings and wrappers with those abstractions for popular optimized platform libraries.
@ -67,16 +68,19 @@ feedback are also welcome.
## Performance ## Performance
Calculation of performance is one of the major goals of KMath in the future, but in some cases it is impossible to achieve both Calculation of performance is one of the major goals of KMath in the future, but in some cases it is impossible to
achieve both
performance and flexibility. performance and flexibility.
We expect to focus on creating a convenient universal API first and then work on increasing performance for specific We expect to focus on creating a convenient universal API first and then work on increasing performance for specific
cases. We expect the worst KMath benchmarks will perform better than native Python, but worse than optimized cases. We expect the worst KMath benchmarks will perform better than native Python, but worse than optimized
native/SciPy (mostly due to boxing operations on primitive numbers). The best performance of optimized parts could be better than SciPy. native/SciPy (mostly due to boxing operations on primitive numbers). The best performance of optimized parts could be
better than SciPy.
## Requirements ## Requirements
KMath currently relies on JDK 11 for compilation and execution of Kotlin-JVM part. We recommend using GraalVM-CE or Oracle GraalVM for execution to get better performance. KMath currently relies on JDK 11 for compilation and execution of Kotlin-JVM part. We recommend using GraalVM-CE or
Oracle GraalVM for execution to get better performance.
### Repositories ### Repositories
@ -99,4 +103,7 @@ dependencies {
## Contributing ## Contributing
The project requires a lot of additional work. The most important thing we need is feedback about what features are The project requires a lot of additional work. The most important thing we need is feedback about what features are
required the most. Feel free to create feature requests. We are also welcome to code contributions, especially in issues marked with [good first issue](hhttps://github.com/SciProgCentre/kmath/issues?q=is%3Aissue+is%3Aopen+label%3A%22good+first+issue%22) label. required the most. Feel free to create feature requests. We are also welcome to code contributions, especially in issues
marked
with [good first issue](hhttps://github.com/SciProgCentre/kmath/issues?q=is%3Aissue+is%3Aopen+label%3A%22good+first+issue%22)
label.

View File

@ -6,6 +6,7 @@
package space.kscience.kmath.expressions package space.kscience.kmath.expressions
import space.kscience.kmath.UnstableKMathAPI import space.kscience.kmath.UnstableKMathAPI
// Only kmath-core is needed. // Only kmath-core is needed.
// Let's declare some variables // Let's declare some variables

View File

@ -8,7 +8,6 @@ package space.kscience.kmath.operations
import space.kscience.kmath.commons.linear.CMLinearSpace import space.kscience.kmath.commons.linear.CMLinearSpace
import space.kscience.kmath.linear.matrix import space.kscience.kmath.linear.matrix
import space.kscience.kmath.nd.Float64BufferND import space.kscience.kmath.nd.Float64BufferND
import space.kscience.kmath.nd.ShapeND
import space.kscience.kmath.nd.Structure2D import space.kscience.kmath.nd.Structure2D
import space.kscience.kmath.nd.mutableStructureND import space.kscience.kmath.nd.mutableStructureND
import space.kscience.kmath.nd.ndAlgebra import space.kscience.kmath.nd.ndAlgebra

View File

@ -53,7 +53,10 @@ class StreamDoubleFieldND(override val shape: ShapeND) : FieldND<Double, Float64
return BufferND(strides, array.asBuffer()) return BufferND(strides, array.asBuffer())
} }
override fun mutableStructureND(shape: ShapeND, initializer: DoubleField.(IntArray) -> Double): MutableBufferND<Double> { override fun mutableStructureND(
shape: ShapeND,
initializer: DoubleField.(IntArray) -> Double,
): MutableBufferND<Double> {
val array = IntStream.range(0, strides.linearSize).parallel().mapToDouble { offset -> val array = IntStream.range(0, strides.linearSize).parallel().mapToDouble { offset ->
val index = strides.index(offset) val index = strides.index(offset)
DoubleField.initializer(index) DoubleField.initializer(index)

View File

@ -12,7 +12,7 @@ import space.kscience.kmath.operations.withSize
inline fun <reified R : Any> MutableBuffer.Companion.same( inline fun <reified R : Any> MutableBuffer.Companion.same(
n: Int, n: Int,
value: R value: R,
): MutableBuffer<R> = MutableBuffer(n) { value } ): MutableBuffer<R> = MutableBuffer(n) { value }

View File

@ -51,7 +51,8 @@ fun main() {
val opts = doubleArrayOf(3.0, 10000.0, 1e-6, 1e-6, 1e-6, 1e-6, 1e-2, 11.0, 9.0, 1.0) val opts = doubleArrayOf(3.0, 10000.0, 1e-6, 1e-6, 1e-6, 1e-6, 1e-2, 11.0, 9.0, 1.0)
// val opts = doubleArrayOf(3.0, 10000.0, 1e-6, 1e-6, 1e-6, 1e-6, 1e-3, 11.0, 9.0, 1.0) // val opts = doubleArrayOf(3.0, 10000.0, 1e-6, 1e-6, 1e-6, 1e-6, 1e-3, 11.0, 9.0, 1.0)
val inputData = LMInput(::funcDifficultForLm, val inputData = LMInput(
::funcDifficultForLm,
p_init.as2D(), p_init.as2D(),
t, t,
y_dat, y_dat,
@ -64,7 +65,8 @@ fun main() {
doubleArrayOf(opts[6], opts[7], opts[8]), doubleArrayOf(opts[6], opts[7], opts[8]),
opts[9].toInt(), opts[9].toInt(),
10, 10,
1) 1
)
val result = DoubleTensorAlgebra.levenbergMarquardt(inputData) val result = DoubleTensorAlgebra.levenbergMarquardt(inputData)

View File

@ -18,7 +18,8 @@ import kotlin.math.roundToInt
fun main() { fun main() {
val startedData = getStartDataForFuncEasy() val startedData = getStartDataForFuncEasy()
val inputData = LMInput(::funcEasyForLm, val inputData = LMInput(
::funcEasyForLm,
DoubleTensorAlgebra.ones(ShapeND(intArrayOf(4, 1))).as2D(), DoubleTensorAlgebra.ones(ShapeND(intArrayOf(4, 1))).as2D(),
startedData.t, startedData.t,
startedData.y_dat, startedData.y_dat,
@ -31,7 +32,8 @@ fun main() {
doubleArrayOf(startedData.opts[6], startedData.opts[7], startedData.opts[8]), doubleArrayOf(startedData.opts[6], startedData.opts[7], startedData.opts[8]),
startedData.opts[9].toInt(), startedData.opts[9].toInt(),
10, 10,
startedData.example_number) startedData.example_number
)
val result = DoubleTensorAlgebra.levenbergMarquardt(inputData) val result = DoubleTensorAlgebra.levenbergMarquardt(inputData)

View File

@ -15,6 +15,7 @@ import space.kscience.kmath.tensors.core.DoubleTensorAlgebra
import space.kscience.kmath.tensors.core.LMInput import space.kscience.kmath.tensors.core.LMInput
import space.kscience.kmath.tensors.core.levenbergMarquardt import space.kscience.kmath.tensors.core.levenbergMarquardt
import kotlin.math.roundToInt import kotlin.math.roundToInt
fun main() { fun main() {
val NData = 100 val NData = 100
var t_example = DoubleTensorAlgebra.ones(ShapeND(intArrayOf(NData, 1))).as2D() var t_example = DoubleTensorAlgebra.ones(ShapeND(intArrayOf(NData, 1))).as2D()
@ -49,7 +50,8 @@ fun main() {
p_min = p_min.div(1.0 / 50.0) p_min = p_min.div(1.0 / 50.0)
val opts = doubleArrayOf(3.0, 7000.0, 1e-5, 1e-5, 1e-5, 1e-5, 1e-5, 11.0, 9.0, 1.0) val opts = doubleArrayOf(3.0, 7000.0, 1e-5, 1e-5, 1e-5, 1e-5, 1e-5, 11.0, 9.0, 1.0)
val inputData = LMInput(::funcMiddleForLm, val inputData = LMInput(
::funcMiddleForLm,
p_init.as2D(), p_init.as2D(),
t, t,
y_dat, y_dat,
@ -62,7 +64,8 @@ fun main() {
doubleArrayOf(opts[6], opts[7], opts[8]), doubleArrayOf(opts[6], opts[7], opts[8]),
opts[9].toInt(), opts[9].toInt(),
10, 10,
1) 1
)
val result = DoubleTensorAlgebra.levenbergMarquardt(inputData) val result = DoubleTensorAlgebra.levenbergMarquardt(inputData)

View File

@ -6,18 +6,23 @@
package space.kscience.kmath.tensors.LevenbergMarquardt.StreamingLm package space.kscience.kmath.tensors.LevenbergMarquardt.StreamingLm
import kotlinx.coroutines.delay import kotlinx.coroutines.delay
import kotlinx.coroutines.flow.* import kotlinx.coroutines.flow.Flow
import space.kscience.kmath.nd.* import kotlinx.coroutines.flow.flow
import space.kscience.kmath.nd.MutableStructure2D
import space.kscience.kmath.nd.ShapeND
import space.kscience.kmath.nd.as2D
import space.kscience.kmath.nd.component1
import space.kscience.kmath.tensors.LevenbergMarquardt.StartDataLm import space.kscience.kmath.tensors.LevenbergMarquardt.StartDataLm
import space.kscience.kmath.tensors.core.BroadcastDoubleTensorAlgebra.zeros import space.kscience.kmath.tensors.core.BroadcastDoubleTensorAlgebra.zeros
import space.kscience.kmath.tensors.core.DoubleTensorAlgebra import space.kscience.kmath.tensors.core.DoubleTensorAlgebra
import space.kscience.kmath.tensors.core.LMInput import space.kscience.kmath.tensors.core.LMInput
import space.kscience.kmath.tensors.core.levenbergMarquardt import space.kscience.kmath.tensors.core.levenbergMarquardt
import kotlin.random.Random import kotlin.random.Random
import kotlin.reflect.KFunction3
fun streamLm(lm_func: (MutableStructure2D<Double>, MutableStructure2D<Double>, Int) -> (MutableStructure2D<Double>), fun streamLm(
startData: StartDataLm, launchFrequencyInMs: Long, numberOfLaunches: Int): Flow<MutableStructure2D<Double>> = flow{ lm_func: (MutableStructure2D<Double>, MutableStructure2D<Double>, Int) -> (MutableStructure2D<Double>),
startData: StartDataLm, launchFrequencyInMs: Long, numberOfLaunches: Int,
): Flow<MutableStructure2D<Double>> = flow {
var example_number = startData.example_number var example_number = startData.example_number
var p_init = startData.p_init var p_init = startData.p_init
@ -32,7 +37,8 @@ fun streamLm(lm_func: (MutableStructure2D<Double>, MutableStructure2D<Double>, I
var steps = numberOfLaunches var steps = numberOfLaunches
val isEndless = (steps <= 0) val isEndless = (steps <= 0)
val inputData = LMInput(lm_func, val inputData = LMInput(
lm_func,
p_init, p_init,
t, t,
y_dat, y_dat,
@ -45,7 +51,8 @@ fun streamLm(lm_func: (MutableStructure2D<Double>, MutableStructure2D<Double>, I
doubleArrayOf(opts[6], opts[7], opts[8]), doubleArrayOf(opts[6], opts[7], opts[8]),
opts[9].toInt(), opts[9].toInt(),
10, 10,
example_number) example_number
)
while (isEndless || steps > 0) { while (isEndless || steps > 0) {
val result = DoubleTensorAlgebra.levenbergMarquardt(inputData) val result = DoubleTensorAlgebra.levenbergMarquardt(inputData)

View File

@ -5,8 +5,9 @@
package space.kscience.kmath.tensors.LevenbergMarquardt.StreamingLm package space.kscience.kmath.tensors.LevenbergMarquardt.StreamingLm
import space.kscience.kmath.nd.* import space.kscience.kmath.nd.component1
import space.kscience.kmath.tensors.LevenbergMarquardt.* import space.kscience.kmath.tensors.LevenbergMarquardt.funcDifficultForLm
import space.kscience.kmath.tensors.LevenbergMarquardt.getStartDataForFuncDifficult
import kotlin.math.roundToInt import kotlin.math.roundToInt
suspend fun main() { suspend fun main() {

View File

@ -29,10 +29,14 @@ public data class StartDataLm (
var p_min: MutableStructure2D<Double>, var p_min: MutableStructure2D<Double>,
var p_max: MutableStructure2D<Double>, var p_max: MutableStructure2D<Double>,
var consts: MutableStructure2D<Double>, var consts: MutableStructure2D<Double>,
var opts: DoubleArray var opts: DoubleArray,
) )
fun funcEasyForLm(t: MutableStructure2D<Double>, p: MutableStructure2D<Double>, exampleNumber: Int): MutableStructure2D<Double> { fun funcEasyForLm(
t: MutableStructure2D<Double>,
p: MutableStructure2D<Double>,
exampleNumber: Int,
): MutableStructure2D<Double> {
val m = t.shape.component1() val m = t.shape.component1()
var y_hat = DoubleTensorAlgebra.zeros(ShapeND(intArrayOf(m, 1))) var y_hat = DoubleTensorAlgebra.zeros(ShapeND(intArrayOf(m, 1)))
@ -40,15 +44,13 @@ fun funcEasyForLm(t: MutableStructure2D<Double>, p: MutableStructure2D<Double>,
y_hat = DoubleTensorAlgebra.exp((t.times(-1.0 / p[1, 0]))).times(p[0, 0]) + t.times(p[2, 0]).times( y_hat = DoubleTensorAlgebra.exp((t.times(-1.0 / p[1, 0]))).times(p[0, 0]) + t.times(p[2, 0]).times(
DoubleTensorAlgebra.exp((t.times(-1.0 / p[3, 0]))) DoubleTensorAlgebra.exp((t.times(-1.0 / p[3, 0])))
) )
} } else if (exampleNumber == 2) {
else if (exampleNumber == 2) {
val mt = t.max() val mt = t.max()
y_hat = (t.times(1.0 / mt)).times(p[0, 0]) + y_hat = (t.times(1.0 / mt)).times(p[0, 0]) +
(t.times(1.0 / mt)).pow(2).times(p[1, 0]) + (t.times(1.0 / mt)).pow(2).times(p[1, 0]) +
(t.times(1.0 / mt)).pow(3).times(p[2, 0]) + (t.times(1.0 / mt)).pow(3).times(p[2, 0]) +
(t.times(1.0 / mt)).pow(4).times(p[3, 0]) (t.times(1.0 / mt)).pow(4).times(p[3, 0])
} } else if (exampleNumber == 3) {
else if (exampleNumber == 3) {
y_hat = DoubleTensorAlgebra.exp((t.times(-1.0 / p[1, 0]))) y_hat = DoubleTensorAlgebra.exp((t.times(-1.0 / p[1, 0])))
.times(p[0, 0]) + DoubleTensorAlgebra.sin((t.times(1.0 / p[3, 0]))).times(p[2, 0]) .times(p[0, 0]) + DoubleTensorAlgebra.sin((t.times(1.0 / p[3, 0]))).times(p[2, 0])
} }
@ -56,7 +58,11 @@ fun funcEasyForLm(t: MutableStructure2D<Double>, p: MutableStructure2D<Double>,
return y_hat.as2D() return y_hat.as2D()
} }
fun funcMiddleForLm(t: MutableStructure2D<Double>, p: MutableStructure2D<Double>, exampleNumber: Int): MutableStructure2D<Double> { fun funcMiddleForLm(
t: MutableStructure2D<Double>,
p: MutableStructure2D<Double>,
exampleNumber: Int,
): MutableStructure2D<Double> {
val m = t.shape.component1() val m = t.shape.component1()
var y_hat = DoubleTensorAlgebra.zeros(ShapeND(intArrayOf(m, 1))) var y_hat = DoubleTensorAlgebra.zeros(ShapeND(intArrayOf(m, 1)))
@ -72,7 +78,11 @@ fun funcMiddleForLm(t: MutableStructure2D<Double>, p: MutableStructure2D<Double>
return y_hat.as2D() return y_hat.as2D()
} }
fun funcDifficultForLm(t: MutableStructure2D<Double>, p: MutableStructure2D<Double>, exampleNumber: Int): MutableStructure2D<Double> { fun funcDifficultForLm(
t: MutableStructure2D<Double>,
p: MutableStructure2D<Double>,
exampleNumber: Int,
): MutableStructure2D<Double> {
val m = t.shape.component1() val m = t.shape.component1()
var y_hat = DoubleTensorAlgebra.zeros(ShapeND(intArrayOf(m, 1))) var y_hat = DoubleTensorAlgebra.zeros(ShapeND(intArrayOf(m, 1)))

View File

@ -5,13 +5,10 @@
kotlin.code.style=official kotlin.code.style=official
kotlin.mpp.stability.nowarn=true kotlin.mpp.stability.nowarn=true
kotlin.native.ignoreDisabledTargets=true kotlin.native.ignoreDisabledTargets=true
org.gradle.configureondemand=true org.gradle.configureondemand=true
org.gradle.jvmargs=-Xmx4096m org.gradle.jvmargs=-Xmx4096m
org.gradle.parallel=true org.gradle.parallel=true
org.gradle.workers.max=4 org.gradle.workers.max=4
toolsVersion=0.15.2-kotlin-1.9.22 toolsVersion=0.15.2-kotlin-1.9.22
#kotlin.experimental.tryK2=true #kotlin.experimental.tryK2=true
#kscience.wasm.disabled=true #kscience.wasm.disabled=true

View File

@ -1,5 +1,5 @@
distributionBase=GRADLE_USER_HOME distributionBase=GRADLE_USER_HOME
distributionPath=wrapper/dists distributionPath=wrapper/dists
distributionUrl=https\://services.gradle.org/distributions/gradle-8.6-bin.zip distributionUrl=https\://services.gradle.org/distributions/gradle-8.7-bin.zip
zipStoreBase=GRADLE_USER_HOME zipStoreBase=GRADLE_USER_HOME
zipStorePath=wrapper/dists zipStorePath=wrapper/dists

View File

@ -7,12 +7,12 @@ Extensions to MST API: transformations, dynamic compilation and visualization.
- [mst-js-codegen](src/jsMain/kotlin/space/kscience/kmath/estree/estree.kt) : Dynamic MST to JS compiler - [mst-js-codegen](src/jsMain/kotlin/space/kscience/kmath/estree/estree.kt) : Dynamic MST to JS compiler
- [rendering](src/commonMain/kotlin/space/kscience/kmath/ast/rendering/MathRenderer.kt) : Extendable MST rendering - [rendering](src/commonMain/kotlin/space/kscience/kmath/ast/rendering/MathRenderer.kt) : Extendable MST rendering
## Artifact: ## Artifact:
The Maven coordinates of this project are `space.kscience:kmath-ast:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-ast:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")
@ -26,21 +26,27 @@ dependencies {
## Parsing expressions ## Parsing expressions
In this module there is a parser from human-readable strings like `"x^3-x+3"` (in the more specific [grammar](reference/ArithmeticsEvaluator.g4)) to MST instances. In this module there is a parser from human-readable strings like `"x^3-x+3"` (in the more
specific [grammar](reference/ArithmeticsEvaluator.g4)) to MST instances.
Supported literals: Supported literals:
1. Constants and variables (consist of latin letters, digits and underscores, can't start with digit): `x`, `_Abc2`. 1. Constants and variables (consist of latin letters, digits and underscores, can't start with digit): `x`, `_Abc2`.
2. Numbers: `123`, `1.02`, `1e10`, `1e-10`, `1.0e+3`&mdash;all parsed either as `kotlin.Long` or `kotlin.Double`. 2. Numbers: `123`, `1.02`, `1e10`, `1e-10`, `1.0e+3`&mdash;all parsed either as `kotlin.Long` or `kotlin.Double`.
Supported binary operators (from the highest precedence to the lowest one): Supported binary operators (from the highest precedence to the lowest one):
1. `^` 1. `^`
2. `*`, `/` 2. `*`, `/`
3. `+`, `-` 3. `+`, `-`
Supported unary operator: Supported unary operator:
1. `-`, e.&nbsp;g. `-x` 1. `-`, e.&nbsp;g. `-x`
Arbitrary unary and binary functions are also supported: names consist of latin letters, digits and underscores, can't start with digit. Examples: Arbitrary unary and binary functions are also supported: names consist of latin letters, digits and underscores, can't
start with digit. Examples:
1. `sin(x)` 1. `sin(x)`
2. `add(x, y)` 2. `add(x, y)`
@ -105,12 +111,15 @@ public final class CompiledExpression_-386104628_0 implements DoubleExpression {
} }
``` ```
Setting JVM system property `space.kscience.kmath.ast.dump.generated.classes` to `1` makes the translator dump class files to program's working directory, so they can be reviewed manually. Setting JVM system property `space.kscience.kmath.ast.dump.generated.classes` to `1` makes the translator dump class
files to program's working directory, so they can be reviewed manually.
#### Limitations #### Limitations
- The same classes may be generated and loaded twice, so it is recommended to cache compiled expressions to avoid class loading overhead. - The same classes may be generated and loaded twice, so it is recommended to cache compiled expressions to avoid class
- This API is not supported by non-dynamic JVM implementations like TeaVM or GraalVM Native Image because they may not support class loaders. loading overhead.
- This API is not supported by non-dynamic JVM implementations like TeaVM or GraalVM Native Image because they may not
support class loaders.
### On JS ### On JS
@ -188,7 +197,8 @@ public fun main() {
Result LaTeX: Result LaTeX:
$$\operatorname{exp}\\,\left(\sqrt{x}\right)-\frac{\frac{\operatorname{arcsin}\\,\left(2\\,x\right)}{2\times10^{10}+x^{3}}}{12}+x^{2/3}$$ $$\operatorname{exp}\\,\left(\sqrt{x}\right)-\frac{\frac{\operatorname{arcsin}\\,\left(2\\,x\right)
}{2\times10^{10}+x^{3}}}{12}+x^{2/3}$$
Result MathML (can be used with MathJax or other renderers): Result MathML (can be used with MathJax or other renderers):

View File

@ -8,21 +8,27 @@ ${artifact}
## Parsing expressions ## Parsing expressions
In this module there is a parser from human-readable strings like `"x^3-x+3"` (in the more specific [grammar](reference/ArithmeticsEvaluator.g4)) to MST instances. In this module there is a parser from human-readable strings like `"x^3-x+3"` (in the more
specific [grammar](reference/ArithmeticsEvaluator.g4)) to MST instances.
Supported literals: Supported literals:
1. Constants and variables (consist of latin letters, digits and underscores, can't start with digit): `x`, `_Abc2`. 1. Constants and variables (consist of latin letters, digits and underscores, can't start with digit): `x`, `_Abc2`.
2. Numbers: `123`, `1.02`, `1e10`, `1e-10`, `1.0e+3`&mdash;all parsed either as `kotlin.Long` or `kotlin.Double`. 2. Numbers: `123`, `1.02`, `1e10`, `1e-10`, `1.0e+3`&mdash;all parsed either as `kotlin.Long` or `kotlin.Double`.
Supported binary operators (from the highest precedence to the lowest one): Supported binary operators (from the highest precedence to the lowest one):
1. `^` 1. `^`
2. `*`, `/` 2. `*`, `/`
3. `+`, `-` 3. `+`, `-`
Supported unary operator: Supported unary operator:
1. `-`, e.&nbsp;g. `-x` 1. `-`, e.&nbsp;g. `-x`
Arbitrary unary and binary functions are also supported: names consist of latin letters, digits and underscores, can't start with digit. Examples: Arbitrary unary and binary functions are also supported: names consist of latin letters, digits and underscores, can't
start with digit. Examples:
1. `sin(x)` 1. `sin(x)`
2. `add(x, y)` 2. `add(x, y)`
@ -87,12 +93,15 @@ public final class CompiledExpression_-386104628_0 implements DoubleExpression {
} }
``` ```
Setting JVM system property `space.kscience.kmath.ast.dump.generated.classes` to `1` makes the translator dump class files to program's working directory, so they can be reviewed manually. Setting JVM system property `space.kscience.kmath.ast.dump.generated.classes` to `1` makes the translator dump class
files to program's working directory, so they can be reviewed manually.
#### Limitations #### Limitations
- The same classes may be generated and loaded twice, so it is recommended to cache compiled expressions to avoid class loading overhead. - The same classes may be generated and loaded twice, so it is recommended to cache compiled expressions to avoid class
- This API is not supported by non-dynamic JVM implementations like TeaVM or GraalVM Native Image because they may not support class loaders. loading overhead.
- This API is not supported by non-dynamic JVM implementations like TeaVM or GraalVM Native Image because they may not
support class loaders.
### On JS ### On JS
@ -170,7 +179,8 @@ public fun main() {
Result LaTeX: Result LaTeX:
$$\operatorname{exp}\\,\left(\sqrt{x}\right)-\frac{\frac{\operatorname{arcsin}\\,\left(2\\,x\right)}{2\times10^{10}+x^{3}}}{12}+x^{2/3}$$ $$\operatorname{exp}\\,\left(\sqrt{x}\right)-\frac{\frac{\operatorname{arcsin}\\,\left(2\\,x\right)
}{2\times10^{10}+x^{3}}}{12}+x^{2/3}$$
Result MathML (can be used with MathJax or other renderers): Result MathML (can be used with MathJax or other renderers):

View File

@ -426,11 +426,13 @@ public class InverseTrigonometricOperations(operations: Collection<String>?) : U
* The default instance configured with [TrigonometricOperations.ACOS_OPERATION], * The default instance configured with [TrigonometricOperations.ACOS_OPERATION],
* [TrigonometricOperations.ASIN_OPERATION], [TrigonometricOperations.ATAN_OPERATION]. * [TrigonometricOperations.ASIN_OPERATION], [TrigonometricOperations.ATAN_OPERATION].
*/ */
public val Default: InverseTrigonometricOperations = InverseTrigonometricOperations(setOf( public val Default: InverseTrigonometricOperations = InverseTrigonometricOperations(
setOf(
TrigonometricOperations.ACOS_OPERATION, TrigonometricOperations.ACOS_OPERATION,
TrigonometricOperations.ASIN_OPERATION, TrigonometricOperations.ASIN_OPERATION,
TrigonometricOperations.ATAN_OPERATION, TrigonometricOperations.ATAN_OPERATION,
)) )
)
} }
} }
@ -452,10 +454,12 @@ public class InverseHyperbolicOperations(operations: Collection<String>?) : Unar
* The default instance configured with [ExponentialOperations.ACOSH_OPERATION], * The default instance configured with [ExponentialOperations.ACOSH_OPERATION],
* [ExponentialOperations.ASINH_OPERATION], and [ExponentialOperations.ATANH_OPERATION]. * [ExponentialOperations.ASINH_OPERATION], and [ExponentialOperations.ATANH_OPERATION].
*/ */
public val Default: InverseHyperbolicOperations = InverseHyperbolicOperations(setOf( public val Default: InverseHyperbolicOperations = InverseHyperbolicOperations(
setOf(
ExponentialOperations.ACOSH_OPERATION, ExponentialOperations.ACOSH_OPERATION,
ExponentialOperations.ASINH_OPERATION, ExponentialOperations.ASINH_OPERATION,
ExponentialOperations.ATANH_OPERATION, ExponentialOperations.ATANH_OPERATION,
)) )
)
} }
} }

View File

@ -17,6 +17,7 @@ internal class TestFeatures {
fun printNumeric() { fun printNumeric() {
val num = object : Number() { val num = object : Number() {
override fun toByte(): Byte = throw UnsupportedOperationException() override fun toByte(): Byte = throw UnsupportedOperationException()
// override fun toChar(): Char = throw UnsupportedOperationException() // override fun toChar(): Char = throw UnsupportedOperationException()
override fun toDouble(): Double = throw UnsupportedOperationException() override fun toDouble(): Double = throw UnsupportedOperationException()
override fun toFloat(): Float = throw UnsupportedOperationException() override fun toFloat(): Float = throw UnsupportedOperationException()

View File

@ -81,8 +81,10 @@ internal class TestMathML {
@Test @Test
fun radicalWithIndex() = fun radicalWithIndex() =
testMathML(RadicalWithIndexSyntax("", SymbolSyntax("x"), SymbolSyntax("y")), testMathML(
"<mroot><mrow><mi>y</mi></mrow><mrow><mi>x</mi></mrow></mroot>") RadicalWithIndexSyntax("", SymbolSyntax("x"), SymbolSyntax("y")),
"<mroot><mrow><mi>y</mi></mrow><mrow><mi>x</mi></mrow></mroot>"
)
@Test @Test
fun multiplication() { fun multiplication() {

View File

@ -1656,27 +1656,27 @@ internal open external class Module {
open fun `if`( open fun `if`(
condition: ExpressionRef, condition: ExpressionRef,
ifTrue: ExpressionRef, ifTrue: ExpressionRef,
ifFalse: ExpressionRef = definedExternally ifFalse: ExpressionRef = definedExternally,
): ExpressionRef ): ExpressionRef
open fun loop(label: String, body: ExpressionRef): ExpressionRef open fun loop(label: String, body: ExpressionRef): ExpressionRef
open fun br( open fun br(
label: String, label: String,
condition: ExpressionRef = definedExternally, condition: ExpressionRef = definedExternally,
value: ExpressionRef = definedExternally value: ExpressionRef = definedExternally,
): ExpressionRef ): ExpressionRef
open fun br_if( open fun br_if(
label: String, label: String,
condition: ExpressionRef = definedExternally, condition: ExpressionRef = definedExternally,
value: ExpressionRef = definedExternally value: ExpressionRef = definedExternally,
): ExpressionRef ): ExpressionRef
open fun switch( open fun switch(
labels: Array<String>, labels: Array<String>,
defaultLabel: String, defaultLabel: String,
condition: ExpressionRef, condition: ExpressionRef,
value: ExpressionRef = definedExternally value: ExpressionRef = definedExternally,
): ExpressionRef ): ExpressionRef
open fun call(name: String, operands: Array<ExpressionRef>, returnType: Type): ExpressionRef open fun call(name: String, operands: Array<ExpressionRef>, returnType: Type): ExpressionRef
@ -1685,14 +1685,14 @@ internal open external class Module {
target: ExpressionRef, target: ExpressionRef,
operands: Array<ExpressionRef>, operands: Array<ExpressionRef>,
params: Type, params: Type,
results: Type results: Type,
): ExpressionRef ): ExpressionRef
open fun return_call_indirect( open fun return_call_indirect(
target: ExpressionRef, target: ExpressionRef,
operands: Array<ExpressionRef>, operands: Array<ExpressionRef>,
params: Type, params: Type,
results: Type results: Type,
): ExpressionRef ): ExpressionRef
open var local: `T$2` open var local: `T$2`
@ -1730,7 +1730,7 @@ internal open external class Module {
condition: ExpressionRef, condition: ExpressionRef,
ifTrue: ExpressionRef, ifTrue: ExpressionRef,
ifFalse: ExpressionRef, ifFalse: ExpressionRef,
type: Type = definedExternally type: Type = definedExternally,
): ExpressionRef ): ExpressionRef
open fun drop(value: ExpressionRef): ExpressionRef open fun drop(value: ExpressionRef): ExpressionRef
@ -1754,7 +1754,7 @@ internal open external class Module {
externalModuleName: String, externalModuleName: String,
externalBaseName: String, externalBaseName: String,
params: Type, params: Type,
results: Type results: Type,
) )
open fun addTableImport(internalName: String, externalModuleName: String, externalBaseName: String) open fun addTableImport(internalName: String, externalModuleName: String, externalBaseName: String)
@ -1763,7 +1763,7 @@ internal open external class Module {
internalName: String, internalName: String,
externalModuleName: String, externalModuleName: String,
externalBaseName: String, externalBaseName: String,
globalType: Type globalType: Type,
) )
open fun addEventImport( open fun addEventImport(
@ -1772,7 +1772,7 @@ internal open external class Module {
externalBaseName: String, externalBaseName: String,
attribute: Number, attribute: Number,
params: Type, params: Type,
results: Type results: Type,
) )
open fun addFunctionExport(internalName: String, externalName: String): ExportRef open fun addFunctionExport(internalName: String, externalName: String): ExportRef
@ -1786,7 +1786,7 @@ internal open external class Module {
initial: Number, initial: Number,
maximum: Number, maximum: Number,
funcNames: Array<Number>, funcNames: Array<Number>,
offset: ExpressionRef = definedExternally offset: ExpressionRef = definedExternally,
) )
open fun getFunctionTable(): `T$26` open fun getFunctionTable(): `T$26`
@ -1796,7 +1796,7 @@ internal open external class Module {
exportName: String? = definedExternally, exportName: String? = definedExternally,
segments: Array<MemorySegment>? = definedExternally, segments: Array<MemorySegment>? = definedExternally,
flags: Array<Number>? = definedExternally, flags: Array<Number>? = definedExternally,
shared: Boolean = definedExternally shared: Boolean = definedExternally,
) )
open fun getNumMemorySegments(): Number open fun getNumMemorySegments(): Number
@ -1827,7 +1827,7 @@ internal open external class Module {
expr: ExpressionRef, expr: ExpressionRef,
fileIndex: Number, fileIndex: Number,
lineNumber: Number, lineNumber: Number,
columnNumber: Number columnNumber: Number,
) )
open fun copyExpression(expr: ExpressionRef): ExpressionRef open fun copyExpression(expr: ExpressionRef): ExpressionRef
@ -2231,7 +2231,7 @@ internal open external class Relooper(module: Module) {
from: RelooperBlockRef, from: RelooperBlockRef,
to: RelooperBlockRef, to: RelooperBlockRef,
indexes: Array<Number>, indexes: Array<Number>,
code: ExpressionRef code: ExpressionRef,
) )
open fun renderAndDispose(entry: RelooperBlockRef, labelHelper: Number): ExpressionRef open fun renderAndDispose(entry: RelooperBlockRef, labelHelper: Number): ExpressionRef

View File

@ -30,7 +30,8 @@ internal fun Identifier(name: String) = object : Identifier {
override var name = name override var name = name
} }
internal fun FunctionExpression(id: Identifier?, params: Array<dynamic>, body: BlockStatement) = object : FunctionExpression { internal fun FunctionExpression(id: Identifier?, params: Array<dynamic>, body: BlockStatement) =
object : FunctionExpression {
override var params = params override var params = params
override var type = "FunctionExpression" override var type = "FunctionExpression"
override var id: Identifier? = id override var id: Identifier? = id

View File

@ -91,6 +91,6 @@ internal typealias Extract<T, U> = Any
internal external interface PromiseLike<T> { internal external interface PromiseLike<T> {
fun then( fun then(
onfulfilled: ((value: T) -> Any?)? = definedExternally, onfulfilled: ((value: T) -> Any?)? = definedExternally,
onrejected: ((reason: Any) -> Any?)? = definedExternally onrejected: ((reason: Any) -> Any?)? = definedExternally,
): PromiseLike<dynamic /* TResult1 | TResult2 */> ): PromiseLike<dynamic /* TResult1 | TResult2 */>
} }

View File

@ -15,11 +15,11 @@
package space.kscience.kmath.internal.webassembly package space.kscience.kmath.internal.webassembly
import space.kscience.kmath.internal.tsstdlib.PromiseLike
import org.khronos.webgl.ArrayBuffer import org.khronos.webgl.ArrayBuffer
import org.khronos.webgl.ArrayBufferView import org.khronos.webgl.ArrayBufferView
import org.khronos.webgl.Uint8Array import org.khronos.webgl.Uint8Array
import org.w3c.fetch.Response import org.w3c.fetch.Response
import space.kscience.kmath.internal.tsstdlib.PromiseLike
import kotlin.js.Promise import kotlin.js.Promise
@Suppress("NESTED_CLASS_IN_EXTERNAL_INTERFACE") @Suppress("NESTED_CLASS_IN_EXTERNAL_INTERFACE")

View File

@ -332,7 +332,7 @@ internal sealed class PrimitiveAsmBuilder<T : Number, out E : Expression<T>>(
private fun visitVariables( private fun visitVariables(
node: TypedMst<T>, node: TypedMst<T>,
arrayMode: Boolean, arrayMode: Boolean,
alreadyLoaded: MutableList<Symbol> = mutableListOf() alreadyLoaded: MutableList<Symbol> = mutableListOf(),
): Unit = when (node) { ): Unit = when (node) {
is TypedMst.Variable -> if (node.symbol !in alreadyLoaded) { is TypedMst.Variable -> if (node.symbol !in alreadyLoaded) {
alreadyLoaded += node.symbol alreadyLoaded += node.symbol

View File

@ -8,7 +8,6 @@ package space.kscience.kmath.asm.internal
import org.objectweb.asm.* import org.objectweb.asm.*
import org.objectweb.asm.commons.InstructionAdapter import org.objectweb.asm.commons.InstructionAdapter
import space.kscience.kmath.expressions.Expression import space.kscience.kmath.expressions.Expression
import space.kscience.kmath.expressions.MST
import kotlin.contracts.InvocationKind import kotlin.contracts.InvocationKind
import kotlin.contracts.contract import kotlin.contracts.contract

View File

@ -9,6 +9,7 @@ Commons math binding for kmath
The Maven coordinates of this project are `space.kscience:kmath-commons:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-commons:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -35,11 +35,13 @@ public class CMGaussRuleIntegrator(
range.start, range.start,
range.endInclusive range.endInclusive
) )
GaussRule.LEGENDREHP -> factory.legendreHighPrecision( GaussRule.LEGENDREHP -> factory.legendreHighPrecision(
numpoints, numpoints,
range.start, range.start,
range.endInclusive range.endInclusive
) )
GaussRule.UNIFORM -> GaussIntegrator( GaussRule.UNIFORM -> GaussIntegrator(
getUniformRule( getUniformRule(
range.start, range.start,

View File

@ -48,9 +48,11 @@ public fun CMLinearSpace.inverse(
public fun CMLinearSpace.solver(decomposition: CMDecomposition): LinearSolver<Double> = object : LinearSolver<Double> { public fun CMLinearSpace.solver(decomposition: CMDecomposition): LinearSolver<Double> = object : LinearSolver<Double> {
override fun solve(a: Matrix<Double>, b: Matrix<Double>): Matrix<Double> = solver(a, decomposition).solve(b.toCM().origin).wrap() override fun solve(a: Matrix<Double>, b: Matrix<Double>): Matrix<Double> =
solver(a, decomposition).solve(b.toCM().origin).wrap()
override fun solve(a: Matrix<Double>, b: Point<Double>): Point<Double> = solver(a, decomposition).solve(b.toCM().origin).toPoint() override fun solve(a: Matrix<Double>, b: Point<Double>): Point<Double> =
solver(a, decomposition).solve(b.toCM().origin).toPoint()
override fun inverse(matrix: Matrix<Double>): Matrix<Double> = solver(matrix, decomposition).inverse.wrap() override fun inverse(matrix: Matrix<Double>): Matrix<Double> = solver(matrix, decomposition).inverse.wrap()
} }

View File

@ -5,12 +5,12 @@ Complex and hypercomplex number systems in KMath.
- [complex](src/commonMain/kotlin/space/kscience/kmath/complex/Complex.kt) : Complex numbers operations - [complex](src/commonMain/kotlin/space/kscience/kmath/complex/Complex.kt) : Complex numbers operations
- [quaternion](src/commonMain/kotlin/space/kscience/kmath/complex/Quaternion.kt) : Quaternions and their composition - [quaternion](src/commonMain/kotlin/space/kscience/kmath/complex/Quaternion.kt) : Quaternions and their composition
## Artifact: ## Artifact:
The Maven coordinates of this project are `space.kscience:kmath-complex:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-complex:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -19,7 +19,8 @@ import kotlin.contracts.contract
*/ */
@OptIn(UnstableKMathAPI::class) @OptIn(UnstableKMathAPI::class)
public sealed class ComplexFieldOpsND : BufferedFieldOpsND<Complex, ComplexField>(ComplexField.bufferAlgebra), public sealed class ComplexFieldOpsND : BufferedFieldOpsND<Complex, ComplexField>(ComplexField.bufferAlgebra),
ScaleOperations<StructureND<Complex>>, ExtendedFieldOps<StructureND<Complex>>, PowerOperations<StructureND<Complex>> { ScaleOperations<StructureND<Complex>>, ExtendedFieldOps<StructureND<Complex>>,
PowerOperations<StructureND<Complex>> {
@OptIn(PerformancePitfall::class) @OptIn(PerformancePitfall::class)
override fun StructureND<Complex>.toBufferND(): BufferND<Complex> = when (this) { override fun StructureND<Complex>.toBufferND(): BufferND<Complex> = when (this) {

View File

@ -2,23 +2,28 @@
The core interfaces of KMath. The core interfaces of KMath.
- [algebras](src/commonMain/kotlin/space/kscience/kmath/operations/Algebra.kt) : Algebraic structures like rings, spaces and fields. - [algebras](src/commonMain/kotlin/space/kscience/kmath/operations/Algebra.kt) : Algebraic structures like rings, spaces
- [nd](src/commonMain/kotlin/space/kscience/kmath/structures/StructureND.kt) : Many-dimensional structures and operations on them. and fields.
- [linear](src/commonMain/kotlin/space/kscience/kmath/operations/Algebra.kt) : Basic linear algebra operations (sums, products, etc.), backed by the `Space` API. Advanced linear algebra operations like matrix inversion and LU decomposition. - [nd](src/commonMain/kotlin/space/kscience/kmath/structures/StructureND.kt) : Many-dimensional structures and
operations on them.
- [linear](src/commonMain/kotlin/space/kscience/kmath/operations/Algebra.kt) : Basic linear algebra operations (sums,
products, etc.), backed by the `Space` API. Advanced linear algebra operations like matrix inversion and LU
decomposition.
- [buffers](src/commonMain/kotlin/space/kscience/kmath/structures/Buffers.kt) : One-dimensional structure - [buffers](src/commonMain/kotlin/space/kscience/kmath/structures/Buffers.kt) : One-dimensional structure
- [expressions](src/commonMain/kotlin/space/kscience/kmath/expressions) : By writing a single mathematical expression once, users will be able to apply different types of - [expressions](src/commonMain/kotlin/space/kscience/kmath/expressions) : By writing a single mathematical expression
once, users will be able to apply different types of
objects to the expression by providing a context. Expressions can be used for a wide variety of purposes from high objects to the expression by providing a context. Expressions can be used for a wide variety of purposes from high
performance calculations to code generation. performance calculations to code generation.
- [domains](src/commonMain/kotlin/space/kscience/kmath/domains) : Domains - [domains](src/commonMain/kotlin/space/kscience/kmath/domains) : Domains
- [autodiff](src/commonMain/kotlin/space/kscience/kmath/expressions/SimpleAutoDiff.kt) : Automatic differentiation - [autodiff](src/commonMain/kotlin/space/kscience/kmath/expressions/SimpleAutoDiff.kt) : Automatic differentiation
- [linear.parallel](#) : Parallel implementation for `LinearAlgebra` - [linear.parallel](#) : Parallel implementation for `LinearAlgebra`
## Artifact: ## Artifact:
The Maven coordinates of this project are `space.kscience:kmath-core:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-core:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -27,7 +27,7 @@ public interface XYErrorColumnarData<T, out X : T, out Y : T> : XYColumnarData<T
public companion object { public companion object {
public fun <T, X : T, Y : T> of( public fun <T, X : T, Y : T> of(
x: Buffer<X>, y: Buffer<Y>, yErr: Buffer<Y> x: Buffer<X>, y: Buffer<Y>, yErr: Buffer<Y>,
): XYErrorColumnarData<T, X, Y> { ): XYErrorColumnarData<T, X, Y> {
require(x.size == y.size) { "Buffer size mismatch. x buffer size is ${x.size}, y buffer size is ${y.size}" } require(x.size == y.size) { "Buffer size mismatch. x buffer size is ${x.size}, y buffer size is ${y.size}" }
require(y.size == yErr.size) { "Buffer size mismatch. y buffer size is ${x.size}, yErr buffer size is ${y.size}" } require(y.size == yErr.size) { "Buffer size mismatch. y buffer size is ${x.size}, yErr buffer size is ${y.size}" }

View File

@ -58,6 +58,7 @@ public fun <T> MST.interpret(algebra: Algebra<T>, arguments: Map<Symbol, T>): T
this.operation, this.operation,
algebra.number(this.value.value), algebra.number(this.value.value),
) )
else -> algebra.unaryOperationFunction(this.operation)(this.value.interpret(algebra, arguments)) else -> algebra.unaryOperationFunction(this.operation)(this.value.interpret(algebra, arguments))
} }

View File

@ -18,12 +18,11 @@ public sealed class Int16RingOpsND : BufferedRingOpsND<Short, Int16Ring>(Int16Ri
@OptIn(UnstableKMathAPI::class) @OptIn(UnstableKMathAPI::class)
public class Int16RingND( public class Int16RingND(
override val shape: ShapeND override val shape: ShapeND,
) : Int16RingOpsND(), RingND<Short, Int16Ring>, NumbersAddOps<StructureND<Short>> { ) : Int16RingOpsND(), RingND<Short, Int16Ring>, NumbersAddOps<StructureND<Short>> {
override fun number(value: Number): BufferND<Short> { override fun number(value: Number): BufferND<Short> {
val short val short = value.toShort() // minimize conversions
= value.toShort() // minimize conversions
return structureND(shape) { short } return structureND(shape) { short }
} }
} }

View File

@ -35,7 +35,7 @@ public sealed class IntRingOpsND : BufferedRingOpsND<Int, Int32Ring>(Int32Ring.b
@OptIn(UnstableKMathAPI::class) @OptIn(UnstableKMathAPI::class)
public class IntRingND( public class IntRingND(
override val shape: ShapeND override val shape: ShapeND,
) : IntRingOpsND(), RingND<Int, Int32Ring>, NumbersAddOps<StructureND<Int>> { ) : IntRingOpsND(), RingND<Int, Int32Ring>, NumbersAddOps<StructureND<Int>> {
override fun number(value: Number): BufferND<Int> { override fun number(value: Number): BufferND<Int> {

View File

@ -14,13 +14,13 @@ import kotlin.jvm.JvmName
public fun <T, A : Algebra<T>> AlgebraND<T, A>.structureND( public fun <T, A : Algebra<T>> AlgebraND<T, A>.structureND(
shapeFirst: Int, shapeFirst: Int,
vararg shapeRest: Int, vararg shapeRest: Int,
initializer: A.(IntArray) -> T initializer: A.(IntArray) -> T,
): StructureND<T> = structureND(ShapeND(shapeFirst, *shapeRest), initializer) ): StructureND<T> = structureND(ShapeND(shapeFirst, *shapeRest), initializer)
public fun <T, A : Algebra<T>> AlgebraND<T, A>.mutableStructureND( public fun <T, A : Algebra<T>> AlgebraND<T, A>.mutableStructureND(
shapeFirst: Int, shapeFirst: Int,
vararg shapeRest: Int, vararg shapeRest: Int,
initializer: A.(IntArray) -> T initializer: A.(IntArray) -> T,
): MutableStructureND<T> = mutableStructureND(ShapeND(shapeFirst, *shapeRest), initializer) ): MutableStructureND<T> = mutableStructureND(ShapeND(shapeFirst, *shapeRest), initializer)
public fun <T, A : Group<T>> AlgebraND<T, A>.zero(shape: ShapeND): StructureND<T> = structureND(shape) { zero } public fun <T, A : Group<T>> AlgebraND<T, A>.zero(shape: ShapeND): StructureND<T> = structureND(shape) { zero }

View File

@ -454,10 +454,12 @@ public fun String.parseBigInteger(): BigInt? {
sign = +1 sign = +1
1 1
} }
'-' -> { '-' -> {
sign = -1 sign = -1
1 1
} }
else -> { else -> {
sign = +1 sign = +1
0 0

View File

@ -99,7 +99,8 @@ public fun <T> Iterable<T>.sumWith(group: Group<T>): T = group.sum(this)
* @param group tha algebra that provides addition * @param group tha algebra that provides addition
* @param extractor the (inline) lambda function to extract value * @param extractor the (inline) lambda function to extract value
*/ */
public inline fun <T, R> Iterable<T>.sumWithGroupOf(group: Group<R>, extractor: (T) -> R): R = this.fold(group.zero) { left: R, right: T -> public inline fun <T, R> Iterable<T>.sumWithGroupOf(group: Group<R>, extractor: (T) -> R): R =
this.fold(group.zero) { left: R, right: T ->
group.add(left, extractor(right)) group.add(left, extractor(right))
} }

View File

@ -55,7 +55,7 @@ public fun FlaggedBuffer<*>.isMissing(index: Int): Boolean = hasFlag(index, Valu
*/ */
public class FlaggedDoubleBuffer( public class FlaggedDoubleBuffer(
public val values: DoubleArray, public val values: DoubleArray,
public val flags: ByteArray public val flags: ByteArray,
) : FlaggedBuffer<Double?>, Buffer<Double?> { ) : FlaggedBuffer<Double?>, Buffer<Double?> {
init { init {

View File

@ -37,7 +37,8 @@ public typealias FloatBuffer = Float32Buffer
* The function [init] is called for each array element sequentially starting from the first one. * The function [init] is called for each array element sequentially starting from the first one.
* It should return the value for a buffer element given its index. * It should return the value for a buffer element given its index.
*/ */
public inline fun Float32Buffer(size: Int, init: (Int) -> Float): Float32Buffer = Float32Buffer(FloatArray(size) { init(it) }) public inline fun Float32Buffer(size: Int, init: (Int) -> Float): Float32Buffer =
Float32Buffer(FloatArray(size) { init(it) })
/** /**
* Returns a new [Float32Buffer] of given elements. * Returns a new [Float32Buffer] of given elements.

View File

@ -74,7 +74,8 @@ public fun <T> Buffer<T>.copy(bufferFactory: BufferFactory<T>): Buffer<T> =if(th
/** /**
* Returns a mutable shallow copy of the buffer. * Returns a mutable shallow copy of the buffer.
*/ */
public fun <T> Buffer<T>.mutableCopy(bufferFactory: MutableBufferFactory<T>): MutableBuffer<T> =if(this is ArrayBuffer){ public fun <T> Buffer<T>.mutableCopy(bufferFactory: MutableBufferFactory<T>): MutableBuffer<T> =
if (this is ArrayBuffer) {
ArrayBuffer(array.copyOf()) ArrayBuffer(array.copyOf())
} else { } else {
bufferFactory(size, ::get) bufferFactory(size, ::get)

View File

@ -67,7 +67,11 @@ class PermSortTest {
assertContentEquals(expected, permutations.map { platforms[it] }, "PermSort using custom ascending comparator") assertContentEquals(expected, permutations.map { platforms[it] }, "PermSort using custom ascending comparator")
permutations = platforms.indicesSortedWith(compareByDescending { it.name.length }) permutations = platforms.indicesSortedWith(compareByDescending { it.name.length })
assertContentEquals(expected.reversed(), permutations.map { platforms[it] }, "PermSort using custom descending comparator") assertContentEquals(
expected.reversed(),
permutations.map { platforms[it] },
"PermSort using custom descending comparator"
)
} }
private fun testPermutation(bufferSize: Int) { private fun testPermutation(bufferSize: Int) {

View File

@ -8,4 +8,5 @@ package space.kscience.kmath.operations
/** /**
* Check if number is an integer * Check if number is an integer
*/ */
public actual fun Number.isInteger(): Boolean = (this is Int) || (this is Long) || (this is Short) || (this.toDouble() % 1 == 0.0) public actual fun Number.isInteger(): Boolean =
(this is Int) || (this is Long) || (this is Short) || (this.toDouble() % 1 == 0.0)

View File

@ -33,7 +33,8 @@ public fun <T> MutableBuffer.Companion.parallel(
typeOf<Double>() -> IntStream.range(0, size).parallel().mapToDouble { initializer(it) as Float64 }.toArray() typeOf<Double>() -> IntStream.range(0, size).parallel().mapToDouble { initializer(it) as Float64 }.toArray()
.asBuffer() as MutableBuffer<T> .asBuffer() as MutableBuffer<T>
//TODO add unsigned types //TODO add unsigned types
else -> IntStream.range(0, size).parallel().mapToObj { initializer(it) }.collect(Collectors.toList<T>()).asMutableBuffer() else -> IntStream.range(0, size).parallel().mapToObj { initializer(it) }.collect(Collectors.toList<T>())
.asMutableBuffer()
} }
public class ParallelBufferFactory<T>(override val type: SafeType<T>) : MutableBufferFactory<T> { public class ParallelBufferFactory<T>(override val type: SafeType<T>) : MutableBufferFactory<T> {

View File

@ -8,4 +8,5 @@ package space.kscience.kmath.operations
/** /**
* Check if number is an integer * Check if number is an integer
*/ */
public actual fun Number.isInteger(): Boolean = (this is Int) || (this is Long) || (this is Short) || (this.toDouble() % 1 == 0.0) public actual fun Number.isInteger(): Boolean =
(this is Int) || (this is Long) || (this is Short) || (this.toDouble() % 1 == 0.0)

View File

@ -1,7 +1,5 @@
# Module kmath-coroutines # Module kmath-coroutines
## Usage ## Usage
## Artifact: ## Artifact:
@ -9,6 +7,7 @@
The Maven coordinates of this project are `space.kscience:kmath-coroutines:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-coroutines:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -25,6 +25,7 @@ public class LazyStructureND<out T>(
} }
public suspend fun await(index: IntArray): T = async(index).await() public suspend fun await(index: IntArray): T = async(index).await()
@PerformancePitfall @PerformancePitfall
override operator fun get(index: IntArray): T = runBlocking { async(index).await() } override operator fun get(index: IntArray): T = runBlocking { async(index).await() }

View File

@ -9,6 +9,7 @@ A proof of concept module for adding type-safe dimensions to structures
The Maven coordinates of this project are `space.kscience:kmath-dimensions:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-dimensions:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -6,12 +6,12 @@ EJML based linear algebra implementation.
- [ejml-matrix](src/main/kotlin/space/kscience/kmath/ejml/EjmlMatrix.kt) : Matrix implementation. - [ejml-matrix](src/main/kotlin/space/kscience/kmath/ejml/EjmlMatrix.kt) : Matrix implementation.
- [ejml-linear-space](src/main/kotlin/space/kscience/kmath/ejml/EjmlLinearSpace.kt) : LinearSpace implementations. - [ejml-linear-space](src/main/kotlin/space/kscience/kmath/ejml/EjmlLinearSpace.kt) : LinearSpace implementations.
## Artifact: ## Artifact:
The Maven coordinates of this project are `space.kscience:kmath-ejml:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-ejml:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -682,7 +682,9 @@ public object EjmlLinearSpaceDSCC : EjmlLinearSpace<Double, Float64Field, DMatri
Determinant -> CommonOps_DSCC.det(origin) Determinant -> CommonOps_DSCC.det(origin)
QR -> object : QRDecomposition<Double> { QR -> object : QRDecomposition<Double> {
val ejmlQr by lazy { DecompositionFactory_DSCC.qr(FillReducing.NONE).apply { decompose(origin.copy()) } } val ejmlQr by lazy {
DecompositionFactory_DSCC.qr(FillReducing.NONE).apply { decompose(origin.copy()) }
}
override val q: Matrix<Double> get() = ejmlQr.getQ(null, false).wrapMatrix() override val q: Matrix<Double> get() = ejmlQr.getQ(null, false).wrapMatrix()
override val r: Matrix<Double> get() = ejmlQr.getR(null, false).wrapMatrix() override val r: Matrix<Double> get() = ejmlQr.getR(null, false).wrapMatrix()
} }
@ -903,7 +905,9 @@ public object EjmlLinearSpaceFSCC : EjmlLinearSpace<Float, Float32Field, FMatrix
Determinant -> CommonOps_FSCC.det(origin) Determinant -> CommonOps_FSCC.det(origin)
QR -> object : QRDecomposition<Float32> { QR -> object : QRDecomposition<Float32> {
val ejmlQr by lazy { DecompositionFactory_FSCC.qr(FillReducing.NONE).apply { decompose(origin.copy()) } } val ejmlQr by lazy {
DecompositionFactory_FSCC.qr(FillReducing.NONE).apply { decompose(origin.copy()) }
}
override val q: Matrix<Float32> get() = ejmlQr.getQ(null, false).wrapMatrix() override val q: Matrix<Float32> get() = ejmlQr.getQ(null, false).wrapMatrix()
override val r: Matrix<Float32> get() = ejmlQr.getR(null, false).wrapMatrix() override val r: Matrix<Float32> get() = ejmlQr.getR(null, false).wrapMatrix()
} }

View File

@ -2,16 +2,18 @@
Specialization of KMath APIs for Double numbers. Specialization of KMath APIs for Double numbers.
- [DoubleVector](src/commonMain/kotlin/space/kscience/kmath/real/DoubleVector.kt) : Numpy-like operations for Buffers/Points - [DoubleVector](src/commonMain/kotlin/space/kscience/kmath/real/DoubleVector.kt) : Numpy-like operations for
- [DoubleMatrix](src/commonMain/kotlin/space/kscience/kmath/real/DoubleMatrix.kt) : Numpy-like operations for 2d real structures Buffers/Points
- [DoubleMatrix](src/commonMain/kotlin/space/kscience/kmath/real/DoubleMatrix.kt) : Numpy-like operations for 2d real
structures
- [grids](src/commonMain/kotlin/space/kscience/kmath/structures/grids.kt) : Uniform grid generators - [grids](src/commonMain/kotlin/space/kscience/kmath/structures/grids.kt) : Uniform grid generators
## Artifact: ## Artifact:
The Maven coordinates of this project are `space.kscience:kmath-for-real:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-for-real:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -4,16 +4,18 @@ Functions and interpolations.
- [piecewise](src/commonMain/kotlin/space/kscience/kmath/functions/Piecewise.kt) : Piecewise functions. - [piecewise](src/commonMain/kotlin/space/kscience/kmath/functions/Piecewise.kt) : Piecewise functions.
- [polynomials](src/commonMain/kotlin/space/kscience/kmath/functions/Polynomial.kt) : Polynomial functions. - [polynomials](src/commonMain/kotlin/space/kscience/kmath/functions/Polynomial.kt) : Polynomial functions.
- [linear interpolation](src/commonMain/kotlin/space/kscience/kmath/interpolation/LinearInterpolator.kt) : Linear XY interpolator. - [linear interpolation](src/commonMain/kotlin/space/kscience/kmath/interpolation/LinearInterpolator.kt) : Linear XY
- [spline interpolation](src/commonMain/kotlin/space/kscience/kmath/interpolation/SplineInterpolator.kt) : Cubic spline XY interpolator. interpolator.
- [spline interpolation](src/commonMain/kotlin/space/kscience/kmath/interpolation/SplineInterpolator.kt) : Cubic spline
XY interpolator.
- [integration](#) : Univariate and multivariate quadratures - [integration](#) : Univariate and multivariate quadratures
## Artifact: ## Artifact:
The Maven coordinates of this project are `space.kscience:kmath-functions:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-functions:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -8,7 +8,8 @@
package space.kscience.kmath.functions package space.kscience.kmath.functions
import space.kscience.kmath.functions.testUtils.* import space.kscience.kmath.functions.testUtils.*
import kotlin.test.* import kotlin.test.Test
import kotlin.test.assertEquals
class PolynomialTest { class PolynomialTest {
@ -52,6 +53,7 @@ class PolynomialTest {
) )
} }
} }
@Test @Test
fun test_Polynomial_Constant_minus() { fun test_Polynomial_Constant_minus() {
RationalField.polynomialSpace { RationalField.polynomialSpace {
@ -92,6 +94,7 @@ class PolynomialTest {
) )
} }
} }
@Test @Test
fun test_Polynomial_Constant_times() { fun test_Polynomial_Constant_times() {
IntModuloRing(35).polynomialSpace { IntModuloRing(35).polynomialSpace {
@ -107,6 +110,7 @@ class PolynomialTest {
) )
} }
} }
@Test @Test
fun test_Constant_Polynomial_plus() { fun test_Constant_Polynomial_plus() {
RationalField.polynomialSpace { RationalField.polynomialSpace {
@ -147,6 +151,7 @@ class PolynomialTest {
) )
} }
} }
@Test @Test
fun test_Constant_Polynomial_minus() { fun test_Constant_Polynomial_minus() {
RationalField.polynomialSpace { RationalField.polynomialSpace {
@ -187,6 +192,7 @@ class PolynomialTest {
) )
} }
} }
@Test @Test
fun test_Constant_Polynomial_times() { fun test_Constant_Polynomial_times() {
IntModuloRing(35).polynomialSpace { IntModuloRing(35).polynomialSpace {
@ -202,6 +208,7 @@ class PolynomialTest {
) )
} }
} }
@Test @Test
fun test_Polynomial_unaryMinus() { fun test_Polynomial_unaryMinus() {
RationalField.polynomialSpace { RationalField.polynomialSpace {
@ -217,6 +224,7 @@ class PolynomialTest {
) )
} }
} }
@Test @Test
fun test_Polynomial_Polynomial_plus() { fun test_Polynomial_Polynomial_plus() {
RationalField.polynomialSpace { RationalField.polynomialSpace {
@ -250,6 +258,7 @@ class PolynomialTest {
) )
} }
} }
@Test @Test
fun test_Polynomial_Polynomial_minus() { fun test_Polynomial_Polynomial_minus() {
RationalField.polynomialSpace { RationalField.polynomialSpace {
@ -283,6 +292,7 @@ class PolynomialTest {
) )
} }
} }
@Test @Test
fun test_Polynomial_Polynomial_times() { fun test_Polynomial_Polynomial_times() {
IntModuloRing(35).polynomialSpace { IntModuloRing(35).polynomialSpace {

View File

@ -53,6 +53,7 @@ class PolynomialUtilTest {
"test 5" "test 5"
) )
} }
@Test @Test
fun test_Polynomial_value_Constant() { fun test_Polynomial_value_Constant() {
assertEquals( assertEquals(
@ -85,6 +86,7 @@ class PolynomialUtilTest {
"test 5" "test 5"
) )
} }
@Test @Test
fun test_Polynomial_differentiate() { fun test_Polynomial_differentiate() {
assertEquals( assertEquals(
@ -94,20 +96,27 @@ class PolynomialUtilTest {
) )
assertEquals( assertEquals(
Polynomial(Rational(-8, 3), Rational(8, 9), Rational(15, 7), Rational(-20, 9)), Polynomial(Rational(-8, 3), Rational(8, 9), Rational(15, 7), Rational(-20, 9)),
Polynomial(Rational(1, 5), Rational(-8, 3), Rational(4, 9), Rational(5, 7), Rational(-5, 9)).differentiate(RationalField), Polynomial(Rational(1, 5), Rational(-8, 3), Rational(4, 9), Rational(5, 7), Rational(-5, 9)).differentiate(
RationalField
),
"test 2" "test 2"
) )
assertEquals( assertEquals(
Polynomial(Rational(0), Rational(8, 9), Rational(15, 7), Rational(-20, 9)), Polynomial(Rational(0), Rational(8, 9), Rational(15, 7), Rational(-20, 9)),
Polynomial(Rational(0), Rational(0), Rational(4, 9), Rational(5, 7), Rational(-5, 9)).differentiate(RationalField), Polynomial(Rational(0), Rational(0), Rational(4, 9), Rational(5, 7), Rational(-5, 9)).differentiate(
RationalField
),
"test 3" "test 3"
) )
assertEquals( assertEquals(
Polynomial(Rational(-8, 3), Rational(8, 9), Rational(15, 7), Rational(0)), Polynomial(Rational(-8, 3), Rational(8, 9), Rational(15, 7), Rational(0)),
Polynomial(Rational(1, 5), Rational(-8, 3), Rational(4, 9), Rational(5, 7), Rational(0)).differentiate(RationalField), Polynomial(Rational(1, 5), Rational(-8, 3), Rational(4, 9), Rational(5, 7), Rational(0)).differentiate(
RationalField
),
"test 4" "test 4"
) )
} }
@Test @Test
fun test_Polynomial_integrate() { fun test_Polynomial_integrate() {
assertEquals( assertEquals(
@ -117,17 +126,23 @@ class PolynomialUtilTest {
) )
assertEquals( assertEquals(
Polynomial(Rational(0), Rational(1, 5), Rational(-4, 3), Rational(4, 27), Rational(5, 28), Rational(-1, 9)), Polynomial(Rational(0), Rational(1, 5), Rational(-4, 3), Rational(4, 27), Rational(5, 28), Rational(-1, 9)),
Polynomial(Rational(1, 5), Rational(-8, 3), Rational(4, 9), Rational(5, 7), Rational(-5, 9)).integrate(RationalField), Polynomial(Rational(1, 5), Rational(-8, 3), Rational(4, 9), Rational(5, 7), Rational(-5, 9)).integrate(
RationalField
),
"test 2" "test 2"
) )
assertEquals( assertEquals(
Polynomial(Rational(0), Rational(0), Rational(0), Rational(4, 27), Rational(5, 28), Rational(-1, 9)), Polynomial(Rational(0), Rational(0), Rational(0), Rational(4, 27), Rational(5, 28), Rational(-1, 9)),
Polynomial(Rational(0), Rational(0), Rational(4, 9), Rational(5, 7), Rational(-5, 9)).integrate(RationalField), Polynomial(Rational(0), Rational(0), Rational(4, 9), Rational(5, 7), Rational(-5, 9)).integrate(
RationalField
),
"test 3" "test 3"
) )
assertEquals( assertEquals(
Polynomial(Rational(0), Rational(1, 5), Rational(-4, 3), Rational(4, 27), Rational(5, 28), Rational(0)), Polynomial(Rational(0), Rational(1, 5), Rational(-4, 3), Rational(4, 27), Rational(5, 28), Rational(0)),
Polynomial(Rational(1, 5), Rational(-8, 3), Rational(4, 9), Rational(5, 7), Rational(0)).integrate(RationalField), Polynomial(Rational(1, 5), Rational(-8, 3), Rational(4, 9), Rational(5, 7), Rational(0)).integrate(
RationalField
),
"test 4" "test 4"
) )
} }

View File

@ -37,6 +37,7 @@ class IntModulo {
modulus, modulus,
toCheckInput = false toCheckInput = false
) )
operator fun plus(other: IntModulo): IntModulo { operator fun plus(other: IntModulo): IntModulo {
require(modulus == other.modulus) { "can not add two residue different modulo" } require(modulus == other.modulus) { "can not add two residue different modulo" }
return IntModulo( return IntModulo(
@ -45,12 +46,14 @@ class IntModulo {
toCheckInput = false toCheckInput = false
) )
} }
operator fun plus(other: Int): IntModulo = operator fun plus(other: Int): IntModulo =
IntModulo( IntModulo(
(residue + other) % modulus, (residue + other) % modulus,
modulus, modulus,
toCheckInput = false toCheckInput = false
) )
operator fun minus(other: IntModulo): IntModulo { operator fun minus(other: IntModulo): IntModulo {
require(modulus == other.modulus) { "can not subtract two residue different modulo" } require(modulus == other.modulus) { "can not subtract two residue different modulo" }
return IntModulo( return IntModulo(
@ -59,12 +62,14 @@ class IntModulo {
toCheckInput = false toCheckInput = false
) )
} }
operator fun minus(other: Int): IntModulo = operator fun minus(other: Int): IntModulo =
IntModulo( IntModulo(
(residue - other) % modulus, (residue - other) % modulus,
modulus, modulus,
toCheckInput = false toCheckInput = false
) )
operator fun times(other: IntModulo): IntModulo { operator fun times(other: IntModulo): IntModulo {
require(modulus == other.modulus) { "can not multiply two residue different modulo" } require(modulus == other.modulus) { "can not multiply two residue different modulo" }
return IntModulo( return IntModulo(
@ -73,12 +78,14 @@ class IntModulo {
toCheckInput = false toCheckInput = false
) )
} }
operator fun times(other: Int): IntModulo = operator fun times(other: Int): IntModulo =
IntModulo( IntModulo(
(residue * other) % modulus, (residue * other) % modulus,
modulus, modulus,
toCheckInput = false toCheckInput = false
) )
operator fun div(other: IntModulo): IntModulo { operator fun div(other: IntModulo): IntModulo {
require(modulus == other.modulus) { "can not divide two residue different modulo" } require(modulus == other.modulus) { "can not divide two residue different modulo" }
val (reciprocalCandidate, gcdOfOtherResidueAndModulus) = bezoutIdentityWithGCD(other.residue, modulus) val (reciprocalCandidate, gcdOfOtherResidueAndModulus) = bezoutIdentityWithGCD(other.residue, modulus)
@ -89,6 +96,7 @@ class IntModulo {
toCheckInput = false toCheckInput = false
) )
} }
operator fun div(other: Int): IntModulo { operator fun div(other: Int): IntModulo {
val (reciprocalCandidate, gcdOfOtherResidueAndModulus) = bezoutIdentityWithGCD(other, modulus) val (reciprocalCandidate, gcdOfOtherResidueAndModulus) = bezoutIdentityWithGCD(other, modulus)
require(gcdOfOtherResidueAndModulus == 1) { "can not divide to residue that has non-trivial GCD with modulo" } require(gcdOfOtherResidueAndModulus == 1) { "can not divide to residue that has non-trivial GCD with modulo" }
@ -98,6 +106,7 @@ class IntModulo {
toCheckInput = false toCheckInput = false
) )
} }
override fun equals(other: Any?): Boolean = override fun equals(other: Any?): Boolean =
when (other) { when (other) {
is IntModulo -> residue == other.residue && modulus == other.modulus is IntModulo -> residue == other.residue && modulus == other.modulus

View File

@ -11,6 +11,7 @@ import space.kscience.kmath.functions.PolynomialSpace
fun PolynomialSpace<IntModulo, IntModuloRing>.Polynomial(vararg coefs: Int): Polynomial<IntModulo> = fun PolynomialSpace<IntModulo, IntModuloRing>.Polynomial(vararg coefs: Int): Polynomial<IntModulo> =
Polynomial(coefs.map { IntModulo(it, ring.modulus) }) Polynomial(coefs.map { IntModulo(it, ring.modulus) })
fun IntModuloRing.Polynomial(vararg coefs: Int): Polynomial<IntModulo> = fun IntModuloRing.Polynomial(vararg coefs: Int): Polynomial<IntModulo> =
Polynomial(coefs.map { IntModulo(it, modulus) }) Polynomial(coefs.map { IntModulo(it, modulus) })

View File

@ -57,18 +57,21 @@ class Rational {
toCheckInput = false toCheckInput = false
) )
} }
operator fun plus(other: Int): Rational = operator fun plus(other: Int): Rational =
Rational( Rational(
numerator + denominator * other.toLong(), numerator + denominator * other.toLong(),
denominator, denominator,
toCheckInput = false toCheckInput = false
) )
operator fun plus(other: Long): Rational = operator fun plus(other: Long): Rational =
Rational( Rational(
numerator + denominator * other, numerator + denominator * other,
denominator, denominator,
toCheckInput = false toCheckInput = false
) )
operator fun minus(other: Rational): Rational { operator fun minus(other: Rational): Rational {
val denominatorsGcd = gcd(denominator, other.denominator) val denominatorsGcd = gcd(denominator, other.denominator)
val dividedThisDenominator = denominator / denominatorsGcd val dividedThisDenominator = denominator / denominatorsGcd
@ -81,18 +84,21 @@ class Rational {
toCheckInput = false toCheckInput = false
) )
} }
operator fun minus(other: Int): Rational = operator fun minus(other: Int): Rational =
Rational( Rational(
numerator - denominator * other.toLong(), numerator - denominator * other.toLong(),
denominator, denominator,
toCheckInput = false toCheckInput = false
) )
operator fun minus(other: Long): Rational = operator fun minus(other: Long): Rational =
Rational( Rational(
numerator - denominator * other, numerator - denominator * other,
denominator, denominator,
toCheckInput = false toCheckInput = false
) )
operator fun times(other: Rational): Rational { operator fun times(other: Rational): Rational {
val thisDenominatorAndOtherNumeratorGcd = gcd(denominator, other.numerator) val thisDenominatorAndOtherNumeratorGcd = gcd(denominator, other.numerator)
val otherDenominatorAndThisNumeratorGcd = gcd(other.denominator, numerator) val otherDenominatorAndThisNumeratorGcd = gcd(other.denominator, numerator)
@ -102,6 +108,7 @@ class Rational {
toCheckInput = false toCheckInput = false
) )
} }
operator fun times(other: Int): Rational { operator fun times(other: Int): Rational {
val other = other.toLong() val other = other.toLong()
val denominatorAndOtherGcd = gcd(denominator, other) val denominatorAndOtherGcd = gcd(denominator, other)
@ -111,6 +118,7 @@ class Rational {
toCheckInput = false toCheckInput = false
) )
} }
operator fun times(other: Long): Rational { operator fun times(other: Long): Rational {
val denominatorAndOtherGcd = gcd(denominator, other) val denominatorAndOtherGcd = gcd(denominator, other)
return Rational( return Rational(
@ -119,6 +127,7 @@ class Rational {
toCheckInput = false toCheckInput = false
) )
} }
operator fun div(other: Rational): Rational { operator fun div(other: Rational): Rational {
val denominatorsGcd = gcd(denominator, other.denominator) val denominatorsGcd = gcd(denominator, other.denominator)
val numeratorsGcd = gcd(numerator, other.numerator) val numeratorsGcd = gcd(numerator, other.numerator)
@ -127,6 +136,7 @@ class Rational {
(denominator / denominatorsGcd) * (other.numerator / numeratorsGcd) (denominator / denominatorsGcd) * (other.numerator / numeratorsGcd)
) )
} }
operator fun div(other: Int): Rational { operator fun div(other: Int): Rational {
val other = other.toLong() val other = other.toLong()
val numeratorAndOtherGcd = gcd(numerator, other) val numeratorAndOtherGcd = gcd(numerator, other)
@ -136,6 +146,7 @@ class Rational {
toCheckInput = false toCheckInput = false
) )
} }
operator fun div(other: Long): Rational { operator fun div(other: Long): Rational {
val numeratorAndOtherGcd = gcd(numerator, other) val numeratorAndOtherGcd = gcd(numerator, other)
return Rational( return Rational(
@ -144,6 +155,7 @@ class Rational {
toCheckInput = false toCheckInput = false
) )
} }
override fun equals(other: Any?): Boolean = override fun equals(other: Any?): Boolean =
when (other) { when (other) {
is Rational -> numerator == other.numerator && denominator == other.denominator is Rational -> numerator == other.numerator && denominator == other.denominator

View File

@ -14,13 +14,41 @@ internal tailrec fun gcd(a: Long, b: Long): Long = if (a == 0L) abs(b) else gcd(
internal fun bezoutIdentityWithGCD(a: Int, b: Int): BezoutIdentityWithGCD<Int> = internal fun bezoutIdentityWithGCD(a: Int, b: Int): BezoutIdentityWithGCD<Int> =
when { when {
a < 0 && b < 0 -> with(bezoutIdentityWithGCDInternalLogic(-a, -b, 1, 0, 0, 1)) { BezoutIdentityWithGCD(-first, -second, gcd) } a < 0 && b < 0 -> with(bezoutIdentityWithGCDInternalLogic(-a, -b, 1, 0, 0, 1)) {
a < 0 -> with(bezoutIdentityWithGCDInternalLogic(-a, b, 1, 0, 0, 1)) { BezoutIdentityWithGCD(-first, second, gcd) } BezoutIdentityWithGCD(
b < 0 -> with(bezoutIdentityWithGCDInternalLogic(a, -b, 1, 0, 0, 1)) { BezoutIdentityWithGCD(first, -second, gcd) } -first,
-second,
gcd
)
}
a < 0 -> with(bezoutIdentityWithGCDInternalLogic(-a, b, 1, 0, 0, 1)) {
BezoutIdentityWithGCD(
-first,
second,
gcd
)
}
b < 0 -> with(bezoutIdentityWithGCDInternalLogic(a, -b, 1, 0, 0, 1)) {
BezoutIdentityWithGCD(
first,
-second,
gcd
)
}
else -> bezoutIdentityWithGCDInternalLogic(a, b, 1, 0, 0, 1) else -> bezoutIdentityWithGCDInternalLogic(a, b, 1, 0, 0, 1)
} }
internal tailrec fun bezoutIdentityWithGCDInternalLogic(a: Int, b: Int, m1: Int, m2: Int, m3: Int, m4: Int): BezoutIdentityWithGCD<Int> = internal tailrec fun bezoutIdentityWithGCDInternalLogic(
a: Int,
b: Int,
m1: Int,
m2: Int,
m3: Int,
m4: Int,
): BezoutIdentityWithGCD<Int> =
if (b == 0) BezoutIdentityWithGCD(m1, m3, a) if (b == 0) BezoutIdentityWithGCD(m1, m3, a)
else { else {
val quotient = a / b val quotient = a / b

View File

@ -1,7 +1,5 @@
# Module kmath-geometry # Module kmath-geometry
## Usage ## Usage
## Artifact: ## Artifact:
@ -9,6 +7,7 @@
The Maven coordinates of this project are `space.kscience:kmath-geometry:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-geometry:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -27,7 +27,6 @@ public typealias Float64Vector2D = Vector2D<Float64>
public typealias DoubleVector2D = Float64Vector2D public typealias DoubleVector2D = Float64Vector2D
/** /**
* 2D Euclidean space * 2D Euclidean space
*/ */

View File

@ -18,7 +18,7 @@ fun ClosedRange<Double>.generateList(step: Double): List<Double> = generateSeque
fun grid( fun grid(
xRange: ClosedRange<Double>, xRange: ClosedRange<Double>,
yRange: ClosedRange<Double>, yRange: ClosedRange<Double>,
step: Double step: Double,
): List<Pair<Double, Double>> { ): List<Pair<Double, Double>> {
val xs = xRange.generateList(step) val xs = xRange.generateList(step)
val ys = yRange.generateList(step) val ys = yRange.generateList(step)

View File

@ -1,7 +1,5 @@
# Module kmath-histograms # Module kmath-histograms
## Usage ## Usage
## Artifact: ## Artifact:
@ -9,6 +7,7 @@
The Maven coordinates of this project are `space.kscience:kmath-histograms:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-histograms:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -4,12 +4,12 @@ Integration with [Jafama](https://github.com/jeffhain/jafama).
- [jafama-double](src/main/kotlin/space/kscience/kmath/jafama/) : Double ExtendedField implementations based on Jafama - [jafama-double](src/main/kotlin/space/kscience/kmath/jafama/) : Double ExtendedField implementations based on Jafama
## Artifact: ## Artifact:
The Maven coordinates of this project are `space.kscience:kmath-jafama:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-jafama:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")
@ -39,6 +39,7 @@ fun main() {
## Performance ## Performance
According to KMath benchmarks on GraalVM, Jafama functions are slower than JDK math; however, there are indications that on Hotspot Jafama is a bit faster. According to KMath benchmarks on GraalVM, Jafama functions are slower than JDK math; however, there are indications that
on Hotspot Jafama is a bit faster.
> **Can't find appropriate benchmark data. Try generating readme files after running benchmarks**. > **Can't find appropriate benchmark data. Try generating readme files after running benchmarks**.

View File

@ -24,6 +24,7 @@ fun main() {
## Performance ## Performance
According to KMath benchmarks on GraalVM, Jafama functions are slower than JDK math; however, there are indications that on Hotspot Jafama is a bit faster. According to KMath benchmarks on GraalVM, Jafama functions are slower than JDK math; however, there are indications that
on Hotspot Jafama is a bit faster.
${benchmarkJafamaDouble} ${benchmarkJafamaDouble}

View File

@ -1,7 +1,5 @@
# Module kmath-jupyter # Module kmath-jupyter
## Usage ## Usage
## Artifact: ## Artifact:
@ -9,6 +7,7 @@
The Maven coordinates of this project are `space.kscience:kmath-jupyter:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-jupyter:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -2,15 +2,17 @@
[Kotlin∇](https://github.com/breandan/kotlingrad) integration module. [Kotlin∇](https://github.com/breandan/kotlingrad) integration module.
- [differentiable-mst-expression](src/main/kotlin/space/kscience/kmath/kotlingrad/KotlingradExpression.kt) : MST based DifferentiableExpression. - [differentiable-mst-expression](src/main/kotlin/space/kscience/kmath/kotlingrad/KotlingradExpression.kt) : MST based
- [scalars-adapters](src/main/kotlin/space/kscience/kmath/kotlingrad/scalarsAdapters.kt) : Conversions between Kotlin∇'s SFun and MST DifferentiableExpression.
- [scalars-adapters](src/main/kotlin/space/kscience/kmath/kotlingrad/scalarsAdapters.kt) : Conversions between Kotlin∇'s
SFun and MST
## Artifact: ## Artifact:
The Maven coordinates of this project are `space.kscience:kmath-kotlingrad:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-kotlingrad:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -1,7 +1,5 @@
# Module kmath-memory # Module kmath-memory
## Usage ## Usage
## Artifact: ## Artifact:
@ -9,6 +7,7 @@
The Maven coordinates of this project are `space.kscience:kmath-memory:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-memory:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -9,6 +9,7 @@ JetBrains Multik connector
The Maven coordinates of this project are `space.kscience:kmath-multik:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-multik:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -16,15 +16,17 @@ import space.kscience.kmath.operations.Float64Field
import space.kscience.kmath.operations.TrigonometricOperations import space.kscience.kmath.operations.TrigonometricOperations
public class MultikDoubleAlgebra( public class MultikDoubleAlgebra(
multikEngine: Engine multikEngine: Engine,
) : MultikDivisionTensorAlgebra<Double, Float64Field>(multikEngine), ) : MultikDivisionTensorAlgebra<Double, Float64Field>(multikEngine),
TrigonometricOperations<StructureND<Double>>, ExponentialOperations<StructureND<Double>> { TrigonometricOperations<StructureND<Double>>, ExponentialOperations<StructureND<Double>> {
override val elementAlgebra: Float64Field get() = Float64Field override val elementAlgebra: Float64Field get() = Float64Field
override val dataType: DataType get() = DataType.DoubleDataType override val dataType: DataType get() = DataType.DoubleDataType
override fun sin(arg: StructureND<Double>): MultikTensor<Double> = multikMath.mathEx.sin(arg.asMultik().array).wrap() override fun sin(arg: StructureND<Double>): MultikTensor<Double> =
multikMath.mathEx.sin(arg.asMultik().array).wrap()
override fun cos(arg: StructureND<Double>): MultikTensor<Double> = multikMath.mathEx.cos(arg.asMultik().array).wrap() override fun cos(arg: StructureND<Double>): MultikTensor<Double> =
multikMath.mathEx.cos(arg.asMultik().array).wrap()
override fun tan(arg: StructureND<Double>): MultikTensor<Double> = sin(arg) / cos(arg) override fun tan(arg: StructureND<Double>): MultikTensor<Double> = sin(arg) / cos(arg)
@ -37,7 +39,8 @@ public class MultikDoubleAlgebra(
@PerformancePitfall @PerformancePitfall
override fun atan(arg: StructureND<Double>): MultikTensor<Double> = arg.map { atan(it) } override fun atan(arg: StructureND<Double>): MultikTensor<Double> = arg.map { atan(it) }
override fun exp(arg: StructureND<Double>): MultikTensor<Double> = multikMath.mathEx.exp(arg.asMultik().array).wrap() override fun exp(arg: StructureND<Double>): MultikTensor<Double> =
multikMath.mathEx.exp(arg.asMultik().array).wrap()
override fun ln(arg: StructureND<Double>): MultikTensor<Double> = multikMath.mathEx.log(arg.asMultik().array).wrap() override fun ln(arg: StructureND<Double>): MultikTensor<Double> = multikMath.mathEx.log(arg.asMultik().array).wrap()

View File

@ -12,7 +12,7 @@ import org.jetbrains.kotlinx.multik.ndarray.data.DataType
import space.kscience.kmath.operations.Float32Field import space.kscience.kmath.operations.Float32Field
public class MultikFloatAlgebra( public class MultikFloatAlgebra(
multikEngine: Engine multikEngine: Engine,
) : MultikDivisionTensorAlgebra<Float, Float32Field>(multikEngine) { ) : MultikDivisionTensorAlgebra<Float, Float32Field>(multikEngine) {
override val elementAlgebra: Float32Field get() = Float32Field override val elementAlgebra: Float32Field get() = Float32Field
override val dataType: DataType get() = DataType.FloatDataType override val dataType: DataType get() = DataType.FloatDataType

View File

@ -12,7 +12,7 @@ import org.jetbrains.kotlinx.multik.ndarray.data.DataType
import space.kscience.kmath.operations.Int32Ring import space.kscience.kmath.operations.Int32Ring
public class MultikIntAlgebra( public class MultikIntAlgebra(
multikEngine: Engine multikEngine: Engine,
) : MultikTensorAlgebra<Int, Int32Ring>(multikEngine) { ) : MultikTensorAlgebra<Int, Int32Ring>(multikEngine) {
override val elementAlgebra: Int32Ring get() = Int32Ring override val elementAlgebra: Int32Ring get() = Int32Ring
override val dataType: DataType get() = DataType.IntDataType override val dataType: DataType get() = DataType.IntDataType

View File

@ -12,7 +12,7 @@ import org.jetbrains.kotlinx.multik.ndarray.data.DataType
import space.kscience.kmath.operations.Int64Ring import space.kscience.kmath.operations.Int64Ring
public class MultikLongAlgebra( public class MultikLongAlgebra(
multikEngine: Engine multikEngine: Engine,
) : MultikTensorAlgebra<Long, Int64Ring>(multikEngine) { ) : MultikTensorAlgebra<Long, Int64Ring>(multikEngine) {
override val elementAlgebra: Int64Ring get() = Int64Ring override val elementAlgebra: Int64Ring get() = Int64Ring
override val dataType: DataType get() = DataType.LongDataType override val dataType: DataType get() = DataType.LongDataType

View File

@ -12,7 +12,7 @@ import org.jetbrains.kotlinx.multik.ndarray.data.DataType
import space.kscience.kmath.operations.Int16Ring import space.kscience.kmath.operations.Int16Ring
public class MultikShortAlgebra( public class MultikShortAlgebra(
multikEngine: Engine multikEngine: Engine,
) : MultikTensorAlgebra<Short, Int16Ring>(multikEngine) { ) : MultikTensorAlgebra<Short, Int16Ring>(multikEngine) {
override val elementAlgebra: Int16Ring get() = Int16Ring override val elementAlgebra: Int16Ring get() = Int16Ring
override val dataType: DataType get() = DataType.ShortDataType override val dataType: DataType get() = DataType.ShortDataType

View File

@ -251,7 +251,12 @@ public abstract class MultikTensorAlgebra<T, A : Ring<T>>(
TODO("Not implemented for broadcasting") TODO("Not implemented for broadcasting")
} }
override fun diagonalEmbedding(diagonalEntries: StructureND<T>, offset: Int, dim1: Int, dim2: Int): MultikTensor<T> { override fun diagonalEmbedding(
diagonalEntries: StructureND<T>,
offset: Int,
dim1: Int,
dim2: Int,
): MultikTensor<T> {
TODO("Diagonal embedding not implemented") TODO("Diagonal embedding not implemented")
} }

View File

@ -6,12 +6,12 @@ ND4J based implementations of KMath abstractions.
- [nd4jarrayrings](#) : Rings over Nd4jArrayStructure of Int and Long - [nd4jarrayrings](#) : Rings over Nd4jArrayStructure of Int and Long
- [nd4jarrayfields](#) : Fields over Nd4jArrayStructure of Float and Double - [nd4jarrayfields](#) : Fields over Nd4jArrayStructure of Float and Double
## Artifact: ## Artifact:
The Maven coordinates of this project are `space.kscience:kmath-nd4j:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-nd4j:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -225,7 +225,8 @@ public open class DoubleNd4jArrayFieldOps : Nd4jArrayExtendedFieldOps<Double, Fl
public val Float64Field.nd4j: DoubleNd4jArrayFieldOps get() = DoubleNd4jArrayFieldOps public val Float64Field.nd4j: DoubleNd4jArrayFieldOps get() = DoubleNd4jArrayFieldOps
public class DoubleNd4jArrayField(override val shape: ShapeND) : DoubleNd4jArrayFieldOps(), FieldND<Double, Float64Field> public class DoubleNd4jArrayField(override val shape: ShapeND) : DoubleNd4jArrayFieldOps(),
FieldND<Double, Float64Field>
public fun Float64Field.nd4j(shapeFirst: Int, vararg shapeRest: Int): DoubleNd4jArrayField = public fun Float64Field.nd4j(shapeFirst: Int, vararg shapeRest: Int): DoubleNd4jArrayField =
DoubleNd4jArrayField(ShapeND(shapeFirst, * shapeRest)) DoubleNd4jArrayField(ShapeND(shapeFirst, * shapeRest))

View File

@ -144,7 +144,9 @@ public sealed interface Nd4jTensorAlgebra<T : Number, A : Field<T>> : AnalyticTe
override fun atanh(arg: StructureND<T>): Nd4jArrayStructure<T> = Transforms.atanh(arg.ndArray).wrap() override fun atanh(arg: StructureND<T>): Nd4jArrayStructure<T> = Transforms.atanh(arg.ndArray).wrap()
override fun power(arg: StructureND<T>, pow: Number): StructureND<T> = Transforms.pow(arg.ndArray, pow).wrap() override fun power(arg: StructureND<T>, pow: Number): StructureND<T> = Transforms.pow(arg.ndArray, pow).wrap()
override fun ceil(arg: StructureND<T>): Nd4jArrayStructure<T> = Transforms.ceil(arg.ndArray).wrap() override fun ceil(arg: StructureND<T>): Nd4jArrayStructure<T> = Transforms.ceil(arg.ndArray).wrap()
override fun floor(structureND: StructureND<T>): Nd4jArrayStructure<T> = Transforms.floor(structureND.ndArray).wrap() override fun floor(structureND: StructureND<T>): Nd4jArrayStructure<T> =
Transforms.floor(structureND.ndArray).wrap()
override fun std(structureND: StructureND<T>, dim: Int, keepDim: Boolean): Tensor<T> = override fun std(structureND: StructureND<T>, dim: Int, keepDim: Boolean): Tensor<T> =
structureND.ndArray.std(true, keepDim, dim).wrap() structureND.ndArray.std(true, keepDim, dim).wrap()
@ -178,7 +180,10 @@ public object DoubleNd4jTensorAlgebra : Nd4jTensorAlgebra<Double, Float64Field>
override fun INDArray.wrap(): Nd4jArrayStructure<Double> = asDoubleStructure() override fun INDArray.wrap(): Nd4jArrayStructure<Double> = asDoubleStructure()
@OptIn(UnsafeKMathAPI::class) @OptIn(UnsafeKMathAPI::class)
override fun mutableStructureND(shape: ShapeND, initializer: Float64Field.(IntArray) -> Double): Nd4jArrayStructure<Double> { override fun mutableStructureND(
shape: ShapeND,
initializer: Float64Field.(IntArray) -> Double,
): Nd4jArrayStructure<Double> {
val array: INDArray = Nd4j.zeros(*shape.asArray()) val array: INDArray = Nd4j.zeros(*shape.asArray())
val indices = ColumnStrides(shape) val indices = ColumnStrides(shape)
indices.asSequence().forEach { index -> indices.asSequence().forEach { index ->

View File

@ -1,7 +1,5 @@
# Module kmath-optimization # Module kmath-optimization
## Usage ## Usage
## Artifact: ## Artifact:
@ -9,6 +7,7 @@
The Maven coordinates of this project are `space.kscience:kmath-optimization:0.4.0-dev-3`. The Maven coordinates of this project are `space.kscience:kmath-optimization:0.4.0-dev-3`.
**Gradle Kotlin DSL:** **Gradle Kotlin DSL:**
```kotlin ```kotlin
repositories { repositories {
maven("https://repo.kotlin.link") maven("https://repo.kotlin.link")

View File

@ -47,7 +47,7 @@ public class FunctionOptimization<T>(
return result return result
} }
override fun toString(): String = "FunctionOptimization(features=$attributes)" override fun toString(): String = "FunctionOptimization(attributes=$attributes)"
public companion object public companion object
} }

View File

@ -319,10 +319,12 @@ public class QowFit(
* generateErrors. * generateErrors.
*/ */
private fun generateErrors(): Matrix<Double> { private fun generateErrors(): Matrix<Double> {
logger?.log { """ logger?.log {
"""
Starting errors estimation using quasioptimal weights method. The starting weight is: Starting errors estimation using quasioptimal weights method. The starting weight is:
${curWeight.theta} ${curWeight.theta}
""".trimIndent()} """.trimIndent()
}
val curWeight = QoWeight(startingPoint) val curWeight = QoWeight(startingPoint)
val covar = getCovariance(curWeight) val covar = getCovariance(curWeight)

View File

@ -33,7 +33,7 @@ internal class CombinedMinimumBuilder : MinimumBuilder {
seed: MinimumSeed?, seed: MinimumSeed?,
strategy: MnStrategy?, strategy: MnStrategy?,
maxfcn: Int, maxfcn: Int,
toler: Double toler: Double,
): FunctionMinimum { ): FunctionMinimum {
val min: FunctionMinimum = theVMMinimizer.minimize(fcn!!, gc, seed, strategy, maxfcn, toler) val min: FunctionMinimum = theVMMinimizer.minimize(fcn!!, gc, seed, strategy, maxfcn, toler)
if (!min.isValid()) { if (!min.isValid()) {

View File

@ -28,7 +28,7 @@ class ContoursError internal constructor(
points: List<Range>, points: List<Range>,
xmnos: MinosError, xmnos: MinosError,
ymnos: MinosError, ymnos: MinosError,
nfcn: Int nfcn: Int,
) { ) {
private val theNFcn: Int private val theNFcn: Int
private val thePoints: List<Range> = points private val thePoints: List<Range> = points

View File

@ -34,8 +34,10 @@ internal class DavidonErrorUpdator : MinimumErrorUpdator {
var Vupd: MnAlgebraicSymMatrix = var Vupd: MnAlgebraicSymMatrix =
MnUtils.sub(MnUtils.div(MnUtils.outerProduct(dx), delgam), MnUtils.div(MnUtils.outerProduct(vg), gvg)) MnUtils.sub(MnUtils.div(MnUtils.outerProduct(dx), delgam), MnUtils.div(MnUtils.outerProduct(vg), gvg))
if (delgam > gvg) { if (delgam > gvg) {
Vupd = MnUtils.add(Vupd, Vupd = MnUtils.add(
MnUtils.mul(MnUtils.outerProduct(MnUtils.sub(MnUtils.div(dx, delgam), MnUtils.div(vg, gvg))), gvg)) Vupd,
MnUtils.mul(MnUtils.outerProduct(MnUtils.sub(MnUtils.div(dx, delgam), MnUtils.div(vg, gvg))), gvg)
)
} }
val sum_upd: Double = MnUtils.absoluteSumOfElements(Vupd) val sum_upd: Double = MnUtils.absoluteSumOfElements(Vupd)
Vupd = MnUtils.add(Vupd, V0) Vupd = MnUtils.add(Vupd, V0)

View File

@ -46,11 +46,15 @@ class FunctionMinimum {
internal constructor(seed: MinimumSeed, up: Double) { internal constructor(seed: MinimumSeed, up: Double) {
theSeed = seed theSeed = seed
theStates = ArrayList() theStates = ArrayList()
theStates.add(MinimumState(seed.parameters(), theStates.add(
MinimumState(
seed.parameters(),
seed.error(), seed.error(),
seed.gradient(), seed.gradient(),
seed.parameters().fval(), seed.parameters().fval(),
seed.nfcn())) seed.nfcn()
)
)
theErrorDef = up theErrorDef = up
theUserState = MnUserParameterState() theUserState = MnUserParameterState()
} }

View File

@ -97,6 +97,7 @@ class MINUITFitter : Fitter {
when (method) { when (method) {
MINUIT_MINOS, MINUIT_MINIMIZE -> minuit = MINUIT_MINOS, MINUIT_MINIMIZE -> minuit =
MnMinimize(fcn, MINUITUtils.getFitParameters(pars, fitPars), strategy) MnMinimize(fcn, MINUITUtils.getFitParameters(pars, fitPars), strategy)
MINUIT_SIMPLEX -> minuit = MnSimplex(fcn, MINUITUtils.getFitParameters(pars, fitPars), strategy) MINUIT_SIMPLEX -> minuit = MnSimplex(fcn, MINUITUtils.getFitParameters(pars, fitPars), strategy)
else -> minuit = MnMigrad(fcn, MINUITUtils.getFitParameters(pars, fitPars), strategy) else -> minuit = MnMigrad(fcn, MINUITUtils.getFitParameters(pars, fitPars), strategy)
} }

View File

@ -13,10 +13,12 @@ import hep.dataforge.context.*
* @author Darksnake * @author Darksnake
* @version $Id: $Id * @version $Id: $Id
*/ */
@PluginDef(group = "hep.dataforge", @PluginDef(
group = "hep.dataforge",
name = "MINUIT", name = "MINUIT",
dependsOn = ["hep.dataforge:fitting"], dependsOn = ["hep.dataforge:fitting"],
info = "The MINUIT fitter engine for DataForge fitting") info = "The MINUIT fitter engine for DataForge fitting"
)
class MINUITPlugin : BasicPlugin() { class MINUITPlugin : BasicPlugin() {
fun attach(@NotNull context: Context?) { fun attach(@NotNull context: Context?) {
super.attach(context) super.attach(context)

View File

@ -40,6 +40,6 @@ interface MinimumBuilder {
seed: MinimumSeed?, seed: MinimumSeed?,
strategy: MnStrategy?, strategy: MnStrategy?,
maxfcn: Int, maxfcn: Int,
toler: Double toler: Double,
): FunctionMinimum ): FunctionMinimum
} }

View File

@ -254,14 +254,16 @@ abstract class MnApplication {
if (maxfcn == 0) { if (maxfcn == 0) {
maxfcn = 200 + 100 * npar + 5 * npar * npar maxfcn = 200 + 100 * npar + 5 * npar * npar
} }
val min: FunctionMinimum = minimizer().minimize(theFCN, val min: FunctionMinimum = minimizer().minimize(
theFCN,
theState, theState,
theStrategy, theStrategy,
maxfcn, maxfcn,
toler, toler,
theErrorDef, theErrorDef,
useAnalyticalDerivatives, useAnalyticalDerivatives,
checkAnalyticalDerivatives) checkAnalyticalDerivatives
)
theNumCall += min.nfcn() theNumCall += min.nfcn()
theState = min.userState() theState = min.userState()
return min return min

View File

@ -113,9 +113,11 @@ class MnContours(fcn: MultiFunction?, min: FunctionMinimum?, stra: MnStrategy?)
return ContoursError(px, py, result, mex, mey, nfcn) return ContoursError(px, py, result, mex, mey, nfcn)
} }
val ey: Range = mey.range() val ey: Range = mey.range()
val migrad = MnMigrad(theFCN, val migrad = MnMigrad(
theFCN,
theMinimum!!.userState().copy(), theMinimum!!.userState().copy(),
MnStrategy(max(0, theStrategy!!.strategy() - 1))) MnStrategy(max(0, theStrategy!!.strategy() - 1))
)
migrad.fix(px) migrad.fix(px)
migrad.setValue(px, valx + ex.getSecond()) migrad.setValue(px, valx + ex.getSecond())
val exy_up: FunctionMinimum = migrad.minimize() val exy_up: FunctionMinimum = migrad.minimize()
@ -131,9 +133,11 @@ class MnContours(fcn: MultiFunction?, min: FunctionMinimum?, stra: MnStrategy?)
MINUITPlugin.logStatic("MnContours is unable to find lower y value for x parameter $px.") MINUITPlugin.logStatic("MnContours is unable to find lower y value for x parameter $px.")
return ContoursError(px, py, result, mex, mey, nfcn) return ContoursError(px, py, result, mex, mey, nfcn)
} }
val migrad1 = MnMigrad(theFCN, val migrad1 = MnMigrad(
theFCN,
theMinimum!!.userState().copy(), theMinimum!!.userState().copy(),
MnStrategy(max(0, theStrategy!!.strategy() - 1))) MnStrategy(max(0, theStrategy!!.strategy() - 1))
)
migrad1.fix(py) migrad1.fix(py)
migrad1.setValue(py, valy + ey.getSecond()) migrad1.setValue(py, valy + ey.getSecond())
val eyx_up: FunctionMinimum = migrad1.minimize() val eyx_up: FunctionMinimum = migrad1.minimize()

View File

@ -28,7 +28,7 @@ internal class MnFunctionCross(
state: MnUserParameterState, state: MnUserParameterState,
fval: Double, fval: Double,
stra: MnStrategy?, stra: MnStrategy?,
errorDef: Double errorDef: Double,
) { ) {
private val theErrorDef: Double private val theErrorDef: Double
private val theFCN: MultiFunction? private val theFCN: MultiFunction?
@ -249,10 +249,13 @@ internal class MnFunctionCross(
} }
} }
do { do {
val parbol: MnParabola = MnParabolaFactory.create(MnParabolaPoint(alsb[0], flsb[0]), val parbol: MnParabola = MnParabolaFactory.create(
MnParabolaPoint(alsb[0], flsb[0]),
MnParabolaPoint(alsb[1], flsb[1]), MnParabolaPoint(alsb[1], flsb[1]),
MnParabolaPoint( MnParabolaPoint(
alsb[2], flsb[2])) alsb[2], flsb[2]
)
)
val coeff1: Double = parbol.c() val coeff1: Double = parbol.c()
val coeff2: Double = parbol.b() val coeff2: Double = parbol.b()
val coeff3: Double = parbol.a() val coeff3: Double = parbol.a()

Some files were not shown because too many files have changed in this diff Show More