[WIP] change data structure
This commit is contained in:
parent
5e3de70737
commit
5fec0518d4
@ -12,7 +12,7 @@ kscience {
|
|||||||
useCoroutines()
|
useCoroutines()
|
||||||
useSerialization()
|
useSerialization()
|
||||||
commonMain {
|
commonMain {
|
||||||
api(project(":dataforge-meta"))
|
api(projects.dataforgeMeta)
|
||||||
api(spclibs.atomicfu)
|
api(spclibs.atomicfu)
|
||||||
}
|
}
|
||||||
jvmMain{
|
jvmMain{
|
||||||
|
@ -1,9 +1,9 @@
|
|||||||
package space.kscience.dataforge.actions
|
package space.kscience.dataforge.actions
|
||||||
|
|
||||||
|
import kotlinx.coroutines.CoroutineScope
|
||||||
import kotlinx.coroutines.launch
|
import kotlinx.coroutines.launch
|
||||||
import space.kscience.dataforge.data.*
|
import space.kscience.dataforge.data.*
|
||||||
import space.kscience.dataforge.meta.Meta
|
import space.kscience.dataforge.meta.Meta
|
||||||
import space.kscience.dataforge.misc.DFInternal
|
|
||||||
import space.kscience.dataforge.names.Name
|
import space.kscience.dataforge.names.Name
|
||||||
import space.kscience.dataforge.names.startsWith
|
import space.kscience.dataforge.names.startsWith
|
||||||
import kotlin.reflect.KType
|
import kotlin.reflect.KType
|
||||||
@ -19,47 +19,40 @@ internal fun MutableMap<Name, *>.removeWhatStartsWith(name: Name) {
|
|||||||
/**
|
/**
|
||||||
* An action that caches results on-demand and recalculates them on source push
|
* An action that caches results on-demand and recalculates them on source push
|
||||||
*/
|
*/
|
||||||
public abstract class AbstractAction<in T : Any, R : Any>(
|
public abstract class AbstractAction<T : Any, R : Any>(
|
||||||
public val outputType: KType,
|
public val outputType: KType,
|
||||||
) : Action<T, R> {
|
) : Action<T, R> {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Generate initial content of the output
|
* Generate initial content of the output
|
||||||
*/
|
*/
|
||||||
protected abstract fun DataSetBuilder<R>.generate(
|
protected abstract fun DataSink<R>.generate(
|
||||||
data: DataSet<T>,
|
data: DataTree<T>,
|
||||||
meta: Meta,
|
meta: Meta,
|
||||||
)
|
)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Update part of the data set when given [updateKey] is triggered by the source
|
* Update part of the data set using provided data
|
||||||
*/
|
*/
|
||||||
protected open fun DataSourceBuilder<R>.update(
|
protected open fun DataSink<R>.update(
|
||||||
dataSet: DataSet<T>,
|
allData: DataTree<T>,
|
||||||
meta: Meta,
|
meta: Meta,
|
||||||
updateKey: Name,
|
namedData: NamedData<T>,
|
||||||
){
|
){
|
||||||
// By default, recalculate the whole dataset
|
//by default regenerate the whole data set
|
||||||
generate(dataSet, meta)
|
generate(allData,meta)
|
||||||
}
|
}
|
||||||
|
|
||||||
@OptIn(DFInternal::class)
|
|
||||||
override fun execute(
|
override fun execute(
|
||||||
dataSet: DataSet<T>,
|
scope: CoroutineScope,
|
||||||
|
dataSet: DataTree<T>,
|
||||||
meta: Meta,
|
meta: Meta,
|
||||||
): DataSet<R> = if (dataSet is DataSource) {
|
): ObservableDataTree<R> = MutableDataTree<R>(outputType, scope).apply {
|
||||||
DataSource(outputType, dataSet){
|
|
||||||
generate(dataSet, meta)
|
generate(dataSet, meta)
|
||||||
|
scope.launch {
|
||||||
launch {
|
dataSet.updates().collect {
|
||||||
dataSet.updates.collect { name ->
|
update(dataSet, meta, it)
|
||||||
update(dataSet, meta, name)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} else {
|
|
||||||
DataTree<R>(outputType) {
|
|
||||||
generate(dataSet, meta)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
@ -1,19 +1,22 @@
|
|||||||
package space.kscience.dataforge.actions
|
package space.kscience.dataforge.actions
|
||||||
|
|
||||||
import space.kscience.dataforge.data.DataSet
|
import kotlinx.coroutines.CoroutineScope
|
||||||
|
import kotlinx.coroutines.coroutineScope
|
||||||
|
import space.kscience.dataforge.data.DataTree
|
||||||
|
import space.kscience.dataforge.data.ObservableDataTree
|
||||||
import space.kscience.dataforge.meta.Meta
|
import space.kscience.dataforge.meta.Meta
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
import space.kscience.dataforge.misc.DFExperimental
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A simple data transformation on a data node. Actions should avoid doing actual dependency evaluation in [execute].
|
* A simple data transformation on a data node. Actions should avoid doing actual dependency evaluation in [execute].
|
||||||
*/
|
*/
|
||||||
public fun interface Action<in T : Any, out R : Any> {
|
public fun interface Action<T : Any, R : Any> {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Transform the data in the node, producing a new node. By default, it is assumed that all calculations are lazy
|
* Transform the data in the node, producing a new node. By default, it is assumed that all calculations are lazy
|
||||||
* so not actual computation is started at this moment.
|
* so not actual computation is started at this moment.
|
||||||
*/
|
*/
|
||||||
public fun execute(dataSet: DataSet<T>, meta: Meta): DataSet<R>
|
public fun execute(scope: CoroutineScope, dataSet: DataTree<T>, meta: Meta): ObservableDataTree<R>
|
||||||
|
|
||||||
public companion object
|
public companion object
|
||||||
}
|
}
|
||||||
@ -21,18 +24,22 @@ public fun interface Action<in T : Any, out R : Any> {
|
|||||||
/**
|
/**
|
||||||
* A convenience method to transform data using given [action]
|
* A convenience method to transform data using given [action]
|
||||||
*/
|
*/
|
||||||
public fun <T : Any, R : Any> DataSet<T>.transform(action: Action<T, R>, meta: Meta = Meta.EMPTY): DataSet<R> =
|
public fun <T : Any, R : Any> DataTree<T>.transform(
|
||||||
action.execute(this, meta)
|
action: Action<T, R>,
|
||||||
|
scope: CoroutineScope,
|
||||||
|
meta: Meta = Meta.EMPTY,
|
||||||
|
): DataTree<R> = action.execute(scope, this, meta)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Action composition. The result is terminal if one of its parts is terminal
|
* Action composition. The result is terminal if one of its parts is terminal
|
||||||
*/
|
*/
|
||||||
public infix fun <T : Any, I : Any, R : Any> Action<T, I>.then(action: Action<I, R>): Action<T, R> =
|
public infix fun <T : Any, I : Any, R : Any> Action<T, I>.then(action: Action<I, R>): Action<T, R> =
|
||||||
Action<T, R> { dataSet, meta -> action.execute(this@then.execute(dataSet, meta), meta) }
|
Action { scope, dataSet, meta -> action.execute(scope, this@then.execute(scope, dataSet, meta), meta) }
|
||||||
|
|
||||||
@DFExperimental
|
@DFExperimental
|
||||||
public operator fun <T : Any, R : Any> Action<T, R>.invoke(
|
public suspend operator fun <T : Any, R : Any> Action<T, R>.invoke(
|
||||||
dataSet: DataSet<T>,
|
dataSet: DataTree<T>,
|
||||||
meta: Meta = Meta.EMPTY,
|
meta: Meta = Meta.EMPTY,
|
||||||
): DataSet<R> = execute(dataSet, meta)
|
): DataTree<R> = coroutineScope { execute(this, dataSet, meta) }
|
||||||
|
|
||||||
|
|
||||||
|
@ -29,6 +29,7 @@ public class MapActionBuilder<T, R>(
|
|||||||
public var name: Name,
|
public var name: Name,
|
||||||
public var meta: MutableMeta,
|
public var meta: MutableMeta,
|
||||||
public val actionMeta: Meta,
|
public val actionMeta: Meta,
|
||||||
|
public val dataType: KType,
|
||||||
@PublishedApi internal var outputType: KType,
|
@PublishedApi internal var outputType: KType,
|
||||||
) {
|
) {
|
||||||
|
|
||||||
@ -45,19 +46,16 @@ public class MapActionBuilder<T, R>(
|
|||||||
/**
|
/**
|
||||||
* Calculate the result of goal
|
* Calculate the result of goal
|
||||||
*/
|
*/
|
||||||
public inline fun <reified R1 : R> result(noinline f: suspend ActionEnv.(T) -> R1) {
|
public inline fun <reified R1 : R> result(noinline f: suspend ActionEnv.(T) -> R1): Unit = result(typeOf<R1>(), f)
|
||||||
outputType = typeOf<R1>()
|
|
||||||
result = f;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@PublishedApi
|
@PublishedApi
|
||||||
internal class MapAction<in T : Any, R : Any>(
|
internal class MapAction<T : Any, R : Any>(
|
||||||
outputType: KType,
|
outputType: KType,
|
||||||
private val block: MapActionBuilder<T, R>.() -> Unit,
|
private val block: MapActionBuilder<T, R>.() -> Unit,
|
||||||
) : AbstractAction<T, R>(outputType) {
|
) : AbstractAction<T, R>(outputType) {
|
||||||
|
|
||||||
private fun DataSetBuilder<R>.mapOne(name: Name, data: Data<T>, meta: Meta) {
|
private fun DataSink<R>.mapOne(name: Name, data: Data<T>, meta: Meta) {
|
||||||
// Creating a new environment for action using **old** name, old meta and task meta
|
// Creating a new environment for action using **old** name, old meta and task meta
|
||||||
val env = ActionEnv(name, data.meta, meta)
|
val env = ActionEnv(name, data.meta, meta)
|
||||||
|
|
||||||
@ -66,6 +64,7 @@ internal class MapAction<in T : Any, R : Any>(
|
|||||||
name,
|
name,
|
||||||
data.meta.toMutableMeta(), // using data meta
|
data.meta.toMutableMeta(), // using data meta
|
||||||
meta,
|
meta,
|
||||||
|
data.type,
|
||||||
outputType
|
outputType
|
||||||
).apply(block)
|
).apply(block)
|
||||||
|
|
||||||
@ -80,16 +79,15 @@ internal class MapAction<in T : Any, R : Any>(
|
|||||||
builder.result(env, data.await())
|
builder.result(env, data.await())
|
||||||
}
|
}
|
||||||
//setting the data node
|
//setting the data node
|
||||||
data(newName, newData)
|
emit(newName, newData)
|
||||||
}
|
}
|
||||||
|
|
||||||
override fun DataSetBuilder<R>.generate(data: DataSet<T>, meta: Meta) {
|
override fun DataSink<R>.generate(data: DataTree<T>, meta: Meta) {
|
||||||
data.forEach { mapOne(it.name, it.data, meta) }
|
data.forEach { mapOne(it.name, it.data, meta) }
|
||||||
}
|
}
|
||||||
|
|
||||||
override fun DataSourceBuilder<R>.update(dataSet: DataSet<T>, meta: Meta, updateKey: Name) {
|
override fun DataSink<R>.update(allData: DataTree<T>, meta: Meta, namedData: NamedData<T>) {
|
||||||
remove(updateKey)
|
mapOne(namedData.name, namedData.data, namedData.meta)
|
||||||
dataSet[updateKey]?.let { mapOne(updateKey, it, meta) }
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -14,7 +14,7 @@ import kotlin.reflect.typeOf
|
|||||||
|
|
||||||
public class JoinGroup<T : Any, R : Any>(
|
public class JoinGroup<T : Any, R : Any>(
|
||||||
public var name: String,
|
public var name: String,
|
||||||
internal val set: DataSet<T>,
|
internal val set: DataTree<T>,
|
||||||
@PublishedApi internal var outputType: KType,
|
@PublishedApi internal var outputType: KType,
|
||||||
) {
|
) {
|
||||||
|
|
||||||
@ -39,7 +39,7 @@ public class ReduceGroupBuilder<T : Any, R : Any>(
|
|||||||
public val actionMeta: Meta,
|
public val actionMeta: Meta,
|
||||||
private val outputType: KType,
|
private val outputType: KType,
|
||||||
) {
|
) {
|
||||||
private val groupRules: MutableList<(DataSet<T>) -> List<JoinGroup<T, R>>> = ArrayList();
|
private val groupRules: MutableList<(DataTree<T>) -> List<JoinGroup<T, R>>> = ArrayList();
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* introduce grouping by meta value
|
* introduce grouping by meta value
|
||||||
@ -54,12 +54,12 @@ public class ReduceGroupBuilder<T : Any, R : Any>(
|
|||||||
|
|
||||||
public fun group(
|
public fun group(
|
||||||
groupName: String,
|
groupName: String,
|
||||||
predicate: (Name, Meta) -> Boolean,
|
predicate: DataFilter,
|
||||||
action: JoinGroup<T, R>.() -> Unit,
|
action: JoinGroup<T, R>.() -> Unit,
|
||||||
) {
|
) {
|
||||||
groupRules += { source ->
|
groupRules += { source ->
|
||||||
listOf(
|
listOf(
|
||||||
JoinGroup<T, R>(groupName, source.filter(predicate), outputType).apply(action)
|
JoinGroup<T, R>(groupName, source.filterData(predicate), outputType).apply(action)
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -73,7 +73,7 @@ public class ReduceGroupBuilder<T : Any, R : Any>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
internal fun buildGroups(input: DataSet<T>): List<JoinGroup<T, R>> =
|
internal fun buildGroups(input: DataTree<T>): List<JoinGroup<T, R>> =
|
||||||
groupRules.flatMap { it.invoke(input) }
|
groupRules.flatMap { it.invoke(input) }
|
||||||
|
|
||||||
}
|
}
|
||||||
@ -85,7 +85,7 @@ internal class ReduceAction<T : Any, R : Any>(
|
|||||||
) : AbstractAction<T, R>(outputType) {
|
) : AbstractAction<T, R>(outputType) {
|
||||||
//TODO optimize reduction. Currently, the whole action recalculates on push
|
//TODO optimize reduction. Currently, the whole action recalculates on push
|
||||||
|
|
||||||
override fun DataSetBuilder<R>.generate(data: DataSet<T>, meta: Meta) {
|
override fun DataSink<R>.generate(data: DataTree<T>, meta: Meta) {
|
||||||
ReduceGroupBuilder<T, R>(meta, outputType).apply(action).buildGroups(data).forEach { group ->
|
ReduceGroupBuilder<T, R>(meta, outputType).apply(action).buildGroups(data).forEach { group ->
|
||||||
val dataFlow: Map<Name, Data<T>> = group.set.asSequence().fold(HashMap()) { acc, value ->
|
val dataFlow: Map<Name, Data<T>> = group.set.asSequence().fold(HashMap()) { acc, value ->
|
||||||
acc.apply {
|
acc.apply {
|
||||||
@ -103,7 +103,7 @@ internal class ReduceAction<T : Any, R : Any>(
|
|||||||
meta = groupMeta
|
meta = groupMeta
|
||||||
) { group.result.invoke(env, it) }
|
) { group.result.invoke(env, it) }
|
||||||
|
|
||||||
data(env.name, res)
|
emit(env.name, res)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -49,7 +49,7 @@ internal class SplitAction<T : Any, R : Any>(
|
|||||||
private val action: SplitBuilder<T, R>.() -> Unit,
|
private val action: SplitBuilder<T, R>.() -> Unit,
|
||||||
) : AbstractAction<T, R>(outputType) {
|
) : AbstractAction<T, R>(outputType) {
|
||||||
|
|
||||||
private fun DataSetBuilder<R>.splitOne(name: Name, data: Data<T>, meta: Meta) {
|
private fun DataSink<R>.splitOne(name: Name, data: Data<T>, meta: Meta) {
|
||||||
val laminate = Laminate(data.meta, meta)
|
val laminate = Laminate(data.meta, meta)
|
||||||
|
|
||||||
val split = SplitBuilder<T, R>(name, data.meta).apply(action)
|
val split = SplitBuilder<T, R>(name, data.meta).apply(action)
|
||||||
@ -64,7 +64,7 @@ internal class SplitAction<T : Any, R : Any>(
|
|||||||
).apply(rule)
|
).apply(rule)
|
||||||
//data.map<R>(outputType, meta = env.meta) { env.result(it) }.named(fragmentName)
|
//data.map<R>(outputType, meta = env.meta) { env.result(it) }.named(fragmentName)
|
||||||
|
|
||||||
data(
|
emit(
|
||||||
fragmentName,
|
fragmentName,
|
||||||
@Suppress("OPT_IN_USAGE") Data(outputType, meta = env.meta, dependencies = listOf(data)) {
|
@Suppress("OPT_IN_USAGE") Data(outputType, meta = env.meta, dependencies = listOf(data)) {
|
||||||
env.result(data.await())
|
env.result(data.await())
|
||||||
@ -73,13 +73,12 @@ internal class SplitAction<T : Any, R : Any>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
override fun DataSetBuilder<R>.generate(data: DataSet<T>, meta: Meta) {
|
override fun DataSink<R>.generate(data: DataTree<T>, meta: Meta) {
|
||||||
data.forEach { splitOne(it.name, it.data, meta) }
|
data.forEach { splitOne(it.name, it.data, meta) }
|
||||||
}
|
}
|
||||||
|
|
||||||
override fun DataSourceBuilder<R>.update(dataSet: DataSet<T>, meta: Meta, updateKey: Name) {
|
override fun DataSink<R>.update(allData: DataTree<T>, meta: Meta, namedData: NamedData<T>) {
|
||||||
remove(updateKey)
|
splitOne(namedData.name, namedData.data, namedData.meta)
|
||||||
dataSet[updateKey]?.let { splitOne(updateKey, it, meta) }
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -15,7 +15,7 @@ import kotlin.reflect.typeOf
|
|||||||
* A data element characterized by its meta
|
* A data element characterized by its meta
|
||||||
*/
|
*/
|
||||||
@DfType(Data.TYPE)
|
@DfType(Data.TYPE)
|
||||||
public interface Data<out T> : Goal<T>, MetaRepr {
|
public interface Data<T> : Goal<T>, MetaRepr {
|
||||||
/**
|
/**
|
||||||
* Type marker for the data. The type is known before the calculation takes place so it could be checked.
|
* Type marker for the data. The type is known before the calculation takes place so it could be checked.
|
||||||
*/
|
*/
|
||||||
@ -41,7 +41,7 @@ public interface Data<out T> : Goal<T>, MetaRepr {
|
|||||||
*/
|
*/
|
||||||
internal val TYPE_OF_NOTHING: KType = typeOf<Unit>()
|
internal val TYPE_OF_NOTHING: KType = typeOf<Unit>()
|
||||||
|
|
||||||
public inline fun <reified T : Any> static(
|
public inline fun <reified T> static(
|
||||||
value: T,
|
value: T,
|
||||||
meta: Meta = Meta.EMPTY,
|
meta: Meta = Meta.EMPTY,
|
||||||
): Data<T> = StaticData(typeOf<T>(), value, meta)
|
): Data<T> = StaticData(typeOf<T>(), value, meta)
|
||||||
@ -69,37 +69,37 @@ public interface Data<out T> : Goal<T>, MetaRepr {
|
|||||||
* A lazily computed variant of [Data] based on [LazyGoal]
|
* A lazily computed variant of [Data] based on [LazyGoal]
|
||||||
* One must ensure that proper [type] is used so this method should not be used
|
* One must ensure that proper [type] is used so this method should not be used
|
||||||
*/
|
*/
|
||||||
private class LazyData<T : Any>(
|
private class LazyData<T>(
|
||||||
override val type: KType,
|
override val type: KType,
|
||||||
override val meta: Meta = Meta.EMPTY,
|
override val meta: Meta = Meta.EMPTY,
|
||||||
additionalContext: CoroutineContext = EmptyCoroutineContext,
|
additionalContext: CoroutineContext = EmptyCoroutineContext,
|
||||||
dependencies: Collection<Goal<*>> = emptyList(),
|
dependencies: Iterable<Goal<*>> = emptyList(),
|
||||||
block: suspend () -> T,
|
block: suspend () -> T,
|
||||||
) : Data<T>, LazyGoal<T>(additionalContext, dependencies, block)
|
) : Data<T>, LazyGoal<T>(additionalContext, dependencies, block)
|
||||||
|
|
||||||
public class StaticData<T : Any>(
|
public class StaticData<T>(
|
||||||
override val type: KType,
|
override val type: KType,
|
||||||
value: T,
|
value: T,
|
||||||
override val meta: Meta = Meta.EMPTY,
|
override val meta: Meta = Meta.EMPTY,
|
||||||
) : Data<T>, StaticGoal<T>(value)
|
) : Data<T>, StaticGoal<T>(value)
|
||||||
|
|
||||||
@Suppress("FunctionName")
|
@Suppress("FunctionName")
|
||||||
public inline fun <reified T : Any> Data(value: T, meta: Meta = Meta.EMPTY): StaticData<T> =
|
public inline fun <reified T> Data(value: T, meta: Meta = Meta.EMPTY): StaticData<T> =
|
||||||
StaticData(typeOf<T>(), value, meta)
|
StaticData(typeOf<T>(), value, meta)
|
||||||
|
|
||||||
@DFInternal
|
@DFInternal
|
||||||
public fun <T : Any> Data(
|
public fun <T> Data(
|
||||||
type: KType,
|
type: KType,
|
||||||
meta: Meta = Meta.EMPTY,
|
meta: Meta = Meta.EMPTY,
|
||||||
context: CoroutineContext = EmptyCoroutineContext,
|
context: CoroutineContext = EmptyCoroutineContext,
|
||||||
dependencies: Collection<Goal<*>> = emptyList(),
|
dependencies: Iterable<Goal<*>> = emptyList(),
|
||||||
block: suspend () -> T,
|
block: suspend () -> T,
|
||||||
): Data<T> = LazyData(type, meta, context, dependencies, block)
|
): Data<T> = LazyData(type, meta, context, dependencies, block)
|
||||||
|
|
||||||
@OptIn(DFInternal::class)
|
@OptIn(DFInternal::class)
|
||||||
public inline fun <reified T : Any> Data(
|
public inline fun <reified T> Data(
|
||||||
meta: Meta = Meta.EMPTY,
|
meta: Meta = Meta.EMPTY,
|
||||||
context: CoroutineContext = EmptyCoroutineContext,
|
context: CoroutineContext = EmptyCoroutineContext,
|
||||||
dependencies: Collection<Goal<*>> = emptyList(),
|
dependencies: Iterable<Goal<*>> = emptyList(),
|
||||||
noinline block: suspend () -> T,
|
noinline block: suspend () -> T,
|
||||||
): Data<T> = Data(typeOf<T>(), meta, context, dependencies, block)
|
): Data<T> = Data(typeOf<T>(), meta, context, dependencies, block)
|
||||||
|
@ -0,0 +1,89 @@
|
|||||||
|
package space.kscience.dataforge.data
|
||||||
|
|
||||||
|
import kotlinx.coroutines.CoroutineScope
|
||||||
|
import kotlinx.coroutines.flow.Flow
|
||||||
|
import kotlinx.coroutines.flow.filter
|
||||||
|
import space.kscience.dataforge.meta.Meta
|
||||||
|
import space.kscience.dataforge.names.Name
|
||||||
|
import kotlin.reflect.KType
|
||||||
|
|
||||||
|
public fun interface DataFilter {
|
||||||
|
|
||||||
|
public fun accepts(name: Name, meta: Meta, type: KType): Boolean
|
||||||
|
|
||||||
|
public companion object {
|
||||||
|
public val EMPTY: DataFilter = DataFilter { _, _, _ -> true }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun DataFilter.accepts(data: NamedData<*>): Boolean = accepts(data.name, data.meta, data.type)
|
||||||
|
|
||||||
|
public fun <T> Sequence<NamedData<T>>.filterData(predicate: DataFilter): Sequence<NamedData<T>> = filter { data ->
|
||||||
|
predicate.accepts(data)
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun <T> Flow<NamedData<T>>.filterData(predicate: DataFilter): Flow<NamedData<T>> = filter { data ->
|
||||||
|
predicate.accepts(data)
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun <T> DataSource<T>.filterData(
|
||||||
|
predicate: DataFilter,
|
||||||
|
): DataSource<T> = object : DataSource<T> {
|
||||||
|
override val dataType: KType get() = this@filterData.dataType
|
||||||
|
|
||||||
|
override fun read(name: Name): Data<T>? =
|
||||||
|
this@filterData.read(name)?.takeIf { predicate.accepts(name, it.meta, it.type) }
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Stateless filtered [ObservableDataSource]
|
||||||
|
*/
|
||||||
|
public fun <T> ObservableDataSource<T>.filterData(
|
||||||
|
predicate: DataFilter,
|
||||||
|
): ObservableDataSource<T> = object : ObservableDataSource<T> {
|
||||||
|
override fun updates(): Flow<NamedData<T>> = this@filterData.updates().filter { predicate.accepts(it) }
|
||||||
|
|
||||||
|
override val dataType: KType get() = this@filterData.dataType
|
||||||
|
|
||||||
|
override fun read(name: Name): Data<T>? =
|
||||||
|
this@filterData.read(name)?.takeIf { predicate.accepts(name, it.meta, it.type) }
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun <T> GenericDataTree<T, *>.filterData(
|
||||||
|
predicate: DataFilter,
|
||||||
|
): DataTree<T> = asSequence().filterData(predicate).toTree(dataType)
|
||||||
|
|
||||||
|
public fun <T> GenericObservableDataTree<T, *>.filterData(
|
||||||
|
scope: CoroutineScope,
|
||||||
|
predicate: DataFilter,
|
||||||
|
): ObservableDataTree<T> = asSequence().filterData(predicate).toObservableTree(dataType, scope, updates().filterData(predicate))
|
||||||
|
|
||||||
|
|
||||||
|
///**
|
||||||
|
// * Generate a wrapper data set with a given name prefix appended to all names
|
||||||
|
// */
|
||||||
|
//public fun <T : Any> DataTree<T>.withNamePrefix(prefix: Name): DataSet<T> = if (prefix.isEmpty()) {
|
||||||
|
// this
|
||||||
|
//} else object : DataSource<T> {
|
||||||
|
//
|
||||||
|
// override val dataType: KType get() = this@withNamePrefix.dataType
|
||||||
|
//
|
||||||
|
// override val coroutineContext: CoroutineContext
|
||||||
|
// get() = (this@withNamePrefix as? DataSource)?.coroutineContext ?: EmptyCoroutineContext
|
||||||
|
//
|
||||||
|
// override val meta: Meta get() = this@withNamePrefix.meta
|
||||||
|
//
|
||||||
|
//
|
||||||
|
// override fun iterator(): Iterator<NamedData<T>> = iterator {
|
||||||
|
// for (d in this@withNamePrefix) {
|
||||||
|
// yield(d.data.named(prefix + d.name))
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// override fun get(name: Name): Data<T>? =
|
||||||
|
// name.removeFirstOrNull(name)?.let { this@withNamePrefix.get(it) }
|
||||||
|
//
|
||||||
|
// override val updates: Flow<Name> get() = this@withNamePrefix.updates.map { prefix + it }
|
||||||
|
//}
|
||||||
|
//
|
||||||
|
|
@ -1,124 +0,0 @@
|
|||||||
package space.kscience.dataforge.data
|
|
||||||
|
|
||||||
import kotlinx.coroutines.*
|
|
||||||
import kotlinx.coroutines.flow.Flow
|
|
||||||
import kotlinx.coroutines.flow.emptyFlow
|
|
||||||
import kotlinx.coroutines.flow.mapNotNull
|
|
||||||
import space.kscience.dataforge.data.Data.Companion.TYPE_OF_NOTHING
|
|
||||||
import space.kscience.dataforge.meta.Meta
|
|
||||||
import space.kscience.dataforge.names.Name
|
|
||||||
import space.kscience.dataforge.names.asName
|
|
||||||
import space.kscience.dataforge.names.endsWith
|
|
||||||
import space.kscience.dataforge.names.parseAsName
|
|
||||||
import kotlin.reflect.KType
|
|
||||||
|
|
||||||
public interface DataSet<out T : Any> {
|
|
||||||
|
|
||||||
/**
|
|
||||||
* The minimal common ancestor to all data in the node
|
|
||||||
*/
|
|
||||||
public val dataType: KType
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Meta-data associated with this node. If no meta is provided, returns [Meta.EMPTY].
|
|
||||||
*/
|
|
||||||
public val meta: Meta
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Traverse this [DataSet] returning named data instances. The order is not guaranteed.
|
|
||||||
*/
|
|
||||||
public operator fun iterator(): Iterator<NamedData<T>>
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Get data with given name.
|
|
||||||
*/
|
|
||||||
public operator fun get(name: Name): Data<T>?
|
|
||||||
|
|
||||||
public companion object {
|
|
||||||
public val META_KEY: Name = "@meta".asName()
|
|
||||||
|
|
||||||
/**
|
|
||||||
* An empty [DataSet] that suits all types
|
|
||||||
*/
|
|
||||||
public val EMPTY: DataSet<Nothing> = object : DataSet<Nothing> {
|
|
||||||
override val dataType: KType = TYPE_OF_NOTHING
|
|
||||||
override val meta: Meta get() = Meta.EMPTY
|
|
||||||
|
|
||||||
override fun iterator(): Iterator<NamedData<Nothing>> = emptySequence<NamedData<Nothing>>().iterator()
|
|
||||||
|
|
||||||
override fun get(name: Name): Data<Nothing>? = null
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public fun <T : Any> DataSet<T>.asSequence(): Sequence<NamedData<T>> = object : Sequence<NamedData<T>> {
|
|
||||||
override fun iterator(): Iterator<NamedData<T>> = this@asSequence.iterator()
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return a single [Data] in this [DataSet]. Throw error if it is not single.
|
|
||||||
*/
|
|
||||||
public fun <T : Any> DataSet<T>.single(): NamedData<T> = asSequence().single()
|
|
||||||
|
|
||||||
public fun <T : Any> DataSet<T>.asIterable(): Iterable<NamedData<T>> = object : Iterable<NamedData<T>> {
|
|
||||||
override fun iterator(): Iterator<NamedData<T>> = this@asIterable.iterator()
|
|
||||||
}
|
|
||||||
|
|
||||||
public operator fun <T : Any> DataSet<T>.get(name: String): Data<T>? = get(name.parseAsName())
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A [DataSet] with propagated updates.
|
|
||||||
*/
|
|
||||||
public interface DataSource<out T : Any> : DataSet<T>, CoroutineScope {
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A flow of updated item names. Updates are propagated in a form of [Flow] of names of updated nodes.
|
|
||||||
* Those can include new data items and replacement of existing ones. The replaced items could update existing data content
|
|
||||||
* and replace it completely, so they should be pulled again.
|
|
||||||
*
|
|
||||||
*/
|
|
||||||
public val updates: Flow<Name>
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Stop generating updates from this [DataSource]
|
|
||||||
*/
|
|
||||||
public fun close() {
|
|
||||||
coroutineContext[Job]?.cancel()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public val <T : Any> DataSet<T>.updates: Flow<Name> get() = if (this is DataSource) updates else emptyFlow()
|
|
||||||
//
|
|
||||||
///**
|
|
||||||
// * Flow all data nodes with names starting with [branchName]
|
|
||||||
// */
|
|
||||||
//public fun <T : Any> DataSet<T>.children(branchName: Name): Sequence<NamedData<T>> =
|
|
||||||
// this@children.asSequence().filter {
|
|
||||||
// it.name.startsWith(branchName)
|
|
||||||
// }
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Start computation for all goals in data node and return a job for the whole node
|
|
||||||
*/
|
|
||||||
public fun <T : Any> DataSet<T>.startAll(coroutineScope: CoroutineScope): Job = coroutineScope.launch {
|
|
||||||
asIterable().map {
|
|
||||||
it.launch(this@launch)
|
|
||||||
}.joinAll()
|
|
||||||
}
|
|
||||||
|
|
||||||
public suspend fun <T : Any> DataSet<T>.computeAndJoinAll(): Unit = coroutineScope { startAll(this).join() }
|
|
||||||
|
|
||||||
public fun DataSet<*>.toMeta(): Meta = Meta {
|
|
||||||
forEach {
|
|
||||||
if (it.name.endsWith(DataSet.META_KEY)) {
|
|
||||||
set(it.name, it.meta)
|
|
||||||
} else {
|
|
||||||
it.name put {
|
|
||||||
"type" put it.type.toString()
|
|
||||||
"meta" put it.meta
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public val <T : Any> DataSet<T>.updatesWithData: Flow<NamedData<T>> get() = updates.mapNotNull { get(it)?.named(it) }
|
|
@ -1,165 +0,0 @@
|
|||||||
package space.kscience.dataforge.data
|
|
||||||
|
|
||||||
import space.kscience.dataforge.meta.Meta
|
|
||||||
import space.kscience.dataforge.meta.MutableMeta
|
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
|
||||||
import space.kscience.dataforge.names.Name
|
|
||||||
import space.kscience.dataforge.names.isEmpty
|
|
||||||
import space.kscience.dataforge.names.plus
|
|
||||||
import kotlin.reflect.KType
|
|
||||||
|
|
||||||
public interface DataSetBuilder<in T : Any> {
|
|
||||||
public val dataType: KType
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Remove all data items starting with [name]
|
|
||||||
*/
|
|
||||||
public fun remove(name: Name)
|
|
||||||
|
|
||||||
public fun data(name: Name, data: Data<T>?)
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Set a current state of given [dataSet] into a branch [name]. Does not propagate updates
|
|
||||||
*/
|
|
||||||
public fun node(name: Name, dataSet: DataSet<T>) {
|
|
||||||
//remove previous items
|
|
||||||
if (name != Name.EMPTY) {
|
|
||||||
remove(name)
|
|
||||||
}
|
|
||||||
|
|
||||||
//Set new items
|
|
||||||
dataSet.forEach {
|
|
||||||
data(name + it.name, it.data)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Set meta for the given node
|
|
||||||
*/
|
|
||||||
public fun meta(name: Name, meta: Meta)
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Define meta in this [DataSet]
|
|
||||||
*/
|
|
||||||
public fun <T : Any> DataSetBuilder<T>.meta(value: Meta): Unit = meta(Name.EMPTY, value)
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Define meta in this [DataSet]
|
|
||||||
*/
|
|
||||||
public fun <T : Any> DataSetBuilder<T>.meta(mutableMeta: MutableMeta.() -> Unit): Unit = meta(Meta(mutableMeta))
|
|
||||||
|
|
||||||
@PublishedApi
|
|
||||||
internal class SubSetBuilder<in T : Any>(
|
|
||||||
private val parent: DataSetBuilder<T>,
|
|
||||||
private val branch: Name,
|
|
||||||
) : DataSetBuilder<T> {
|
|
||||||
override val dataType: KType get() = parent.dataType
|
|
||||||
|
|
||||||
override fun remove(name: Name) {
|
|
||||||
parent.remove(branch + name)
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun data(name: Name, data: Data<T>?) {
|
|
||||||
parent.data(branch + name, data)
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun node(name: Name, dataSet: DataSet<T>) {
|
|
||||||
parent.node(branch + name, dataSet)
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun meta(name: Name, meta: Meta) {
|
|
||||||
parent.meta(branch + name, meta)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public inline fun <T : Any> DataSetBuilder<T>.node(
|
|
||||||
name: Name,
|
|
||||||
crossinline block: DataSetBuilder<T>.() -> Unit,
|
|
||||||
) {
|
|
||||||
if (name.isEmpty()) block() else SubSetBuilder(this, name).block()
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
public fun <T : Any> DataSetBuilder<T>.data(name: String, value: Data<T>) {
|
|
||||||
data(Name.parse(name), value)
|
|
||||||
}
|
|
||||||
|
|
||||||
public fun <T : Any> DataSetBuilder<T>.node(name: String, set: DataSet<T>) {
|
|
||||||
node(Name.parse(name), set)
|
|
||||||
}
|
|
||||||
|
|
||||||
public inline fun <T : Any> DataSetBuilder<T>.node(
|
|
||||||
name: String,
|
|
||||||
crossinline block: DataSetBuilder<T>.() -> Unit,
|
|
||||||
): Unit = node(Name.parse(name), block)
|
|
||||||
|
|
||||||
public fun <T : Any> DataSetBuilder<T>.set(value: NamedData<T>) {
|
|
||||||
data(value.name, value.data)
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Produce lazy [Data] and emit it into the [DataSetBuilder]
|
|
||||||
*/
|
|
||||||
public inline fun <reified T : Any> DataSetBuilder<T>.produce(
|
|
||||||
name: String,
|
|
||||||
meta: Meta = Meta.EMPTY,
|
|
||||||
noinline producer: suspend () -> T,
|
|
||||||
) {
|
|
||||||
val data = Data(meta, block = producer)
|
|
||||||
data(name, data)
|
|
||||||
}
|
|
||||||
|
|
||||||
public inline fun <reified T : Any> DataSetBuilder<T>.produce(
|
|
||||||
name: Name,
|
|
||||||
meta: Meta = Meta.EMPTY,
|
|
||||||
noinline producer: suspend () -> T,
|
|
||||||
) {
|
|
||||||
val data = Data(meta, block = producer)
|
|
||||||
data(name, data)
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Emit a static data with the fixed value
|
|
||||||
*/
|
|
||||||
public inline fun <reified T : Any> DataSetBuilder<T>.static(
|
|
||||||
name: String,
|
|
||||||
data: T,
|
|
||||||
meta: Meta = Meta.EMPTY,
|
|
||||||
): Unit = data(name, Data.static(data, meta))
|
|
||||||
|
|
||||||
public inline fun <reified T : Any> DataSetBuilder<T>.static(
|
|
||||||
name: Name,
|
|
||||||
data: T,
|
|
||||||
meta: Meta = Meta.EMPTY,
|
|
||||||
): Unit = data(name, Data.static(data, meta))
|
|
||||||
|
|
||||||
public inline fun <reified T : Any> DataSetBuilder<T>.static(
|
|
||||||
name: String,
|
|
||||||
data: T,
|
|
||||||
mutableMeta: MutableMeta.() -> Unit,
|
|
||||||
): Unit = data(Name.parse(name), Data.static(data, Meta(mutableMeta)))
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Update data with given node data and meta with node meta.
|
|
||||||
*/
|
|
||||||
@DFExperimental
|
|
||||||
public fun <T : Any> DataSetBuilder<T>.populateFrom(tree: DataSet<T>): Unit {
|
|
||||||
tree.forEach {
|
|
||||||
//TODO check if the place is occupied
|
|
||||||
data(it.name, it.data)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
//public fun <T : Any> DataSetBuilder<T>.populateFrom(flow: Flow<NamedData<T>>) {
|
|
||||||
// flow.collect {
|
|
||||||
// data(it.name, it.data)
|
|
||||||
// }
|
|
||||||
//}
|
|
||||||
|
|
||||||
public fun <T : Any> DataSetBuilder<T>.populateFrom(sequence: Sequence<NamedData<T>>) {
|
|
||||||
sequence.forEach {
|
|
||||||
data(it.name, it.data)
|
|
||||||
}
|
|
||||||
}
|
|
@ -0,0 +1,271 @@
|
|||||||
|
package space.kscience.dataforge.data
|
||||||
|
|
||||||
|
import kotlinx.coroutines.CoroutineScope
|
||||||
|
import kotlinx.coroutines.flow.*
|
||||||
|
import kotlinx.coroutines.launch
|
||||||
|
import space.kscience.dataforge.meta.Meta
|
||||||
|
import space.kscience.dataforge.misc.DFInternal
|
||||||
|
import space.kscience.dataforge.names.*
|
||||||
|
import kotlin.reflect.KType
|
||||||
|
import kotlin.reflect.typeOf
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A generic data provider
|
||||||
|
*/
|
||||||
|
public interface DataSource<T> {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The minimal common ancestor to all data in the node
|
||||||
|
*/
|
||||||
|
public val dataType: KType
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get data with given name. Or null if it is not present
|
||||||
|
*/
|
||||||
|
public fun read(name: Name): Data<T>?
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A data provider with possible dynamic updates
|
||||||
|
*/
|
||||||
|
public interface ObservableDataSource<T> : DataSource<T> {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Flow updates made to the data
|
||||||
|
*/
|
||||||
|
public fun updates(): Flow<NamedData<T>>
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A tree like structure for data holding
|
||||||
|
*/
|
||||||
|
public interface GenericDataTree<T, out TR : GenericDataTree<T, TR>> : DataSource<T> {
|
||||||
|
public val self: TR
|
||||||
|
|
||||||
|
public val data: Data<T>?
|
||||||
|
public val items: Map<NameToken, TR>
|
||||||
|
|
||||||
|
|
||||||
|
override fun read(name: Name): Data<T>? = when (name.length) {
|
||||||
|
0 -> data
|
||||||
|
else -> items[name.first()]?.read(name.cutFirst())
|
||||||
|
}
|
||||||
|
|
||||||
|
public companion object {
|
||||||
|
private object EmptyDataTree : GenericDataTree<Nothing, EmptyDataTree> {
|
||||||
|
override val self: EmptyDataTree get() = this
|
||||||
|
override val data: Data<Nothing>? = null
|
||||||
|
override val items: Map<NameToken, EmptyDataTree> = emptyMap()
|
||||||
|
override val dataType: KType = typeOf<Unit>()
|
||||||
|
|
||||||
|
override fun read(name: Name): Data<Nothing>? = null
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
public val EMPTY: GenericDataTree<Nothing, *> = EmptyDataTree
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public typealias DataTree<T> = GenericDataTree<T, *>
|
||||||
|
|
||||||
|
/**
|
||||||
|
* An alias for easier access to tree values
|
||||||
|
*/
|
||||||
|
public operator fun <T> DataTree<T>.get(name: Name): Data<T>? = read(name)
|
||||||
|
|
||||||
|
public operator fun <T> DataTree<T>.get(name: String): Data<T>? = read(name.parseAsName())
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Return a sequence of all data items in this tree.
|
||||||
|
* This method does not take updates into account.
|
||||||
|
*/
|
||||||
|
public fun <T> GenericDataTree<T, DataTree<T>>.asSequence(
|
||||||
|
namePrefix: Name = Name.EMPTY,
|
||||||
|
): Sequence<NamedData<T>> = sequence {
|
||||||
|
data?.let { yield(it.named(Name.EMPTY)) }
|
||||||
|
items.forEach { (token, tree) ->
|
||||||
|
yieldAll(tree.asSequence(namePrefix + token))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public val DataTree<*>.meta: Meta? get() = data?.meta
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Provide subtree if it exists
|
||||||
|
*/
|
||||||
|
public tailrec fun <T, TR : GenericDataTree<T, TR>> GenericDataTree<T, TR>.branch(name: Name): TR? =
|
||||||
|
when (name.length) {
|
||||||
|
0 -> self
|
||||||
|
1 -> items[name.first()]
|
||||||
|
else -> items[name.first()]?.branch(name.cutFirst())
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun GenericDataTree<*, *>.isEmpty(): Boolean = data == null && items.isEmpty()
|
||||||
|
|
||||||
|
@PublishedApi
|
||||||
|
internal class FlatDataTree<T>(
|
||||||
|
override val dataType: KType,
|
||||||
|
val dataSet: Map<Name, Data<T>>,
|
||||||
|
val prefix: Name,
|
||||||
|
) : GenericDataTree<T, FlatDataTree<T>> {
|
||||||
|
override val self: FlatDataTree<T> get() = this
|
||||||
|
override val data: Data<T>? get() = dataSet[prefix]
|
||||||
|
override val items: Map<NameToken, FlatDataTree<T>>
|
||||||
|
get() = dataSet.keys
|
||||||
|
.filter { it.startsWith(prefix) && it.length > prefix.length }
|
||||||
|
.map { it.tokens[prefix.length + 1] }
|
||||||
|
.associateWith { FlatDataTree(dataType, dataSet, prefix + it) }
|
||||||
|
|
||||||
|
override fun read(name: Name): Data<T>? = dataSet[prefix + name]
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Represent this flat data map as a [DataTree] without copying it
|
||||||
|
*/
|
||||||
|
public inline fun <reified T> Map<Name, Data<T>>.asTree(): DataTree<T> = FlatDataTree(typeOf<T>(), this, Name.EMPTY)
|
||||||
|
|
||||||
|
internal fun <T> Sequence<NamedData<T>>.toTree(type: KType): DataTree<T> =
|
||||||
|
FlatDataTree(type, associate { it.name to it.data }, Name.EMPTY)
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Collect a sequence of [NamedData] to a [DataTree]
|
||||||
|
*/
|
||||||
|
public inline fun <reified T> Sequence<NamedData<T>>.toTree(): DataTree<T> =
|
||||||
|
FlatDataTree(typeOf<T>(), associate { it.name to it.data }, Name.EMPTY)
|
||||||
|
|
||||||
|
public interface GenericObservableDataTree<T, TR : GenericObservableDataTree<T, TR>> : GenericDataTree<T, TR>,
|
||||||
|
ObservableDataSource<T>
|
||||||
|
|
||||||
|
public typealias ObservableDataTree<T> = GenericObservableDataTree<T, *>
|
||||||
|
|
||||||
|
public fun <T> DataTree<T>.updates(): Flow<NamedData<T>> = if (this is ObservableDataTree<T>) updates() else emptyFlow()
|
||||||
|
|
||||||
|
public fun interface DataSink<T> {
|
||||||
|
public fun emit(name: Name, data: Data<T>?)
|
||||||
|
}
|
||||||
|
|
||||||
|
public class DataTreeBuilder<T>(private val type: KType) : DataSink<T> {
|
||||||
|
private val map = HashMap<Name, Data<T>>()
|
||||||
|
override fun emit(name: Name, data: Data<T>?) {
|
||||||
|
if (data == null) {
|
||||||
|
map.remove(name)
|
||||||
|
} else {
|
||||||
|
map[name] = data
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun build(): DataTree<T> = FlatDataTree(type, map, Name.EMPTY)
|
||||||
|
}
|
||||||
|
|
||||||
|
@DFInternal
|
||||||
|
public inline fun <T> DataTree(
|
||||||
|
dataType: KType,
|
||||||
|
generator: DataSink<T>.() -> Unit,
|
||||||
|
): DataTree<T> = DataTreeBuilder<T>(dataType).apply(generator).build()
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Create and a data tree.
|
||||||
|
*/
|
||||||
|
public inline fun <reified T> DataTree(
|
||||||
|
generator: DataSink<T>.() -> Unit,
|
||||||
|
): DataTree<T> = DataTreeBuilder<T>(typeOf<T>()).apply(generator).build()
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A mutable version of [GenericDataTree]
|
||||||
|
*/
|
||||||
|
public interface MutableDataTree<T> : GenericObservableDataTree<T, MutableDataTree<T>>, DataSink<T> {
|
||||||
|
public val scope: CoroutineScope
|
||||||
|
|
||||||
|
override var data: Data<T>?
|
||||||
|
|
||||||
|
override val items: Map<NameToken, MutableDataTree<T>>
|
||||||
|
|
||||||
|
public operator fun set(token: NameToken, data: Data<T>?)
|
||||||
|
|
||||||
|
override fun emit(name: Name, data: Data<T>?): Unit = set(name, data)
|
||||||
|
}
|
||||||
|
|
||||||
|
public tailrec operator fun <T> MutableDataTree<T>.set(name: Name, data: Data<T>?): Unit {
|
||||||
|
when (name.length) {
|
||||||
|
0 -> this.data = data
|
||||||
|
1 -> set(name.first(), data)
|
||||||
|
else -> items[name.first()]?.set(name.cutFirst(), data)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private class ObservableMutableDataTreeImpl<T>(
|
||||||
|
override val dataType: KType,
|
||||||
|
override val scope: CoroutineScope,
|
||||||
|
) : MutableDataTree<T> {
|
||||||
|
|
||||||
|
private val updates = MutableSharedFlow<NamedData<T>>()
|
||||||
|
|
||||||
|
private val children = HashMap<NameToken, MutableDataTree<T>>()
|
||||||
|
|
||||||
|
override var data: Data<T>? = null
|
||||||
|
set(value) {
|
||||||
|
field = value
|
||||||
|
if (value != null) {
|
||||||
|
scope.launch {
|
||||||
|
updates.emit(value.named(Name.EMPTY))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override val items: Map<NameToken, MutableDataTree<T>> get() = children
|
||||||
|
override val self: MutableDataTree<T> get() = this
|
||||||
|
|
||||||
|
override fun set(token: NameToken, data: Data<T>?) {
|
||||||
|
children.getOrPut(token) {
|
||||||
|
ObservableMutableDataTreeImpl<T>(dataType, scope).also { subTree ->
|
||||||
|
subTree.updates().onEach {
|
||||||
|
updates.emit(it.named(token + it.name))
|
||||||
|
}.launchIn(scope)
|
||||||
|
}
|
||||||
|
}.data = data
|
||||||
|
}
|
||||||
|
|
||||||
|
override fun updates(): Flow<NamedData<T>> = flow {
|
||||||
|
//emit this node updates
|
||||||
|
updates.collect {
|
||||||
|
emit(it)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun <T> MutableDataTree(
|
||||||
|
type: KType,
|
||||||
|
scope: CoroutineScope
|
||||||
|
): MutableDataTree<T> = ObservableMutableDataTreeImpl<T>(type, scope)
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Create and initialize a observable mutable data tree.
|
||||||
|
*/
|
||||||
|
public inline fun <reified T> MutableDataTree(
|
||||||
|
scope: CoroutineScope,
|
||||||
|
generator: MutableDataTree<T>.() -> Unit = {},
|
||||||
|
): MutableDataTree<T> = MutableDataTree<T>(typeOf<T>(), scope).apply { generator() }
|
||||||
|
|
||||||
|
//@DFInternal
|
||||||
|
//public fun <T> ObservableDataTree(
|
||||||
|
// type: KType,
|
||||||
|
// scope: CoroutineScope,
|
||||||
|
// generator: suspend MutableDataTree<T>.() -> Unit = {},
|
||||||
|
//): ObservableDataTree<T> = MutableDataTree<T>(type, scope.coroutineContext).apply(generator)
|
||||||
|
|
||||||
|
public inline fun <reified T> ObservableDataTree(
|
||||||
|
scope: CoroutineScope,
|
||||||
|
generator: MutableDataTree<T>.() -> Unit = {},
|
||||||
|
): ObservableDataTree<T> = MutableDataTree<T>(typeOf<T>(), scope).apply(generator)
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Collect a [Sequence] into an observable tree with additional [updates]
|
||||||
|
*/
|
||||||
|
public fun <T> Sequence<NamedData<T>>.toObservableTree(dataType: KType, scope: CoroutineScope, updates: Flow<NamedData<T>>): ObservableDataTree<T> =
|
||||||
|
MutableDataTree<T>(dataType, scope).apply {
|
||||||
|
emitAll(this@toObservableTree)
|
||||||
|
updates.onEach {
|
||||||
|
emit(it.name, it.data)
|
||||||
|
}.launchIn(scope)
|
||||||
|
}
|
@ -1,119 +0,0 @@
|
|||||||
package space.kscience.dataforge.data
|
|
||||||
|
|
||||||
import space.kscience.dataforge.meta.Meta
|
|
||||||
import space.kscience.dataforge.misc.DFInternal
|
|
||||||
import space.kscience.dataforge.misc.DfType
|
|
||||||
import space.kscience.dataforge.names.*
|
|
||||||
import kotlin.collections.component1
|
|
||||||
import kotlin.collections.component2
|
|
||||||
import kotlin.reflect.KType
|
|
||||||
import kotlin.reflect.typeOf
|
|
||||||
|
|
||||||
public sealed class DataTreeItem<out T : Any> {
|
|
||||||
|
|
||||||
public abstract val meta: Meta
|
|
||||||
|
|
||||||
public class Node<out T : Any>(public val tree: DataTree<T>) : DataTreeItem<T>() {
|
|
||||||
override val meta: Meta get() = tree.meta
|
|
||||||
}
|
|
||||||
|
|
||||||
public class Leaf<out T : Any>(public val data: Data<T>) : DataTreeItem<T>() {
|
|
||||||
override val meta: Meta get() = data.meta
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public val <T : Any> DataTreeItem<T>.type: KType
|
|
||||||
get() = when (this) {
|
|
||||||
is DataTreeItem.Node -> tree.dataType
|
|
||||||
is DataTreeItem.Leaf -> data.type
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A tree-like [DataSet] grouped into the node. All data inside the node must inherit its type
|
|
||||||
*/
|
|
||||||
@DfType(DataTree.TYPE)
|
|
||||||
public interface DataTree<out T : Any> : DataSet<T> {
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Top-level children items of this [DataTree]
|
|
||||||
*/
|
|
||||||
public val items: Map<NameToken, DataTreeItem<T>>
|
|
||||||
|
|
||||||
override val meta: Meta get() = items[META_ITEM_NAME_TOKEN]?.meta ?: Meta.EMPTY
|
|
||||||
|
|
||||||
override fun iterator(): Iterator<NamedData<T>> = iterator {
|
|
||||||
items.forEach { (token, childItem: DataTreeItem<T>) ->
|
|
||||||
if (!token.body.startsWith("@")) {
|
|
||||||
when (childItem) {
|
|
||||||
is DataTreeItem.Leaf -> yield(childItem.data.named(token.asName()))
|
|
||||||
is DataTreeItem.Node -> yieldAll(childItem.tree.asSequence().map { it.named(token + it.name) })
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun get(name: Name): Data<T>? = when (name.length) {
|
|
||||||
0 -> null
|
|
||||||
1 -> items[name.firstOrNull()!!].data
|
|
||||||
else -> items[name.firstOrNull()!!].tree?.get(name.cutFirst())
|
|
||||||
}
|
|
||||||
|
|
||||||
public companion object {
|
|
||||||
public const val TYPE: String = "dataTree"
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A name token used to designate tree node meta
|
|
||||||
*/
|
|
||||||
public val META_ITEM_NAME_TOKEN: NameToken = NameToken("@meta")
|
|
||||||
|
|
||||||
@DFInternal
|
|
||||||
public fun <T : Any> emptyWithType(type: KType, meta: Meta = Meta.EMPTY): DataTree<T> = object : DataTree<T> {
|
|
||||||
override val items: Map<NameToken, DataTreeItem<T>> get() = emptyMap()
|
|
||||||
override val dataType: KType get() = type
|
|
||||||
override val meta: Meta get() = meta
|
|
||||||
}
|
|
||||||
|
|
||||||
@OptIn(DFInternal::class)
|
|
||||||
public inline fun <reified T : Any> empty(meta: Meta = Meta.EMPTY): DataTree<T> =
|
|
||||||
emptyWithType<T>(typeOf<T>(), meta)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public fun <T : Any> DataTree<T>.listChildren(prefix: Name): List<Name> =
|
|
||||||
getItem(prefix).tree?.items?.keys?.map { prefix + it } ?: emptyList()
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Get a [DataTreeItem] with given [name] or null if the item does not exist
|
|
||||||
*/
|
|
||||||
public tailrec fun <T : Any> DataTree<T>.getItem(name: Name): DataTreeItem<T>? = when (name.length) {
|
|
||||||
0 -> DataTreeItem.Node(this)
|
|
||||||
1 -> items[name.firstOrNull()]
|
|
||||||
else -> items[name.firstOrNull()!!].tree?.getItem(name.cutFirst())
|
|
||||||
}
|
|
||||||
|
|
||||||
public val <T : Any> DataTreeItem<T>?.tree: DataTree<T>? get() = (this as? DataTreeItem.Node<T>)?.tree
|
|
||||||
public val <T : Any> DataTreeItem<T>?.data: Data<T>? get() = (this as? DataTreeItem.Leaf<T>)?.data
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A [Sequence] of all children including nodes
|
|
||||||
*/
|
|
||||||
public fun <T : Any> DataTree<T>.traverseItems(): Sequence<Pair<Name, DataTreeItem<T>>> = sequence {
|
|
||||||
items.forEach { (head, item) ->
|
|
||||||
yield(head.asName() to item)
|
|
||||||
if (item is DataTreeItem.Node) {
|
|
||||||
val subSequence = item.tree.traverseItems()
|
|
||||||
.map { (name, data) -> (head.asName() + name) to data }
|
|
||||||
yieldAll(subSequence)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Get a branch of this [DataTree] with a given [branchName].
|
|
||||||
* The difference from similar method for [DataSet] is that internal logic is more simple and the return value is a [DataTree]
|
|
||||||
*/
|
|
||||||
@OptIn(DFInternal::class)
|
|
||||||
public fun <T : Any> DataTree<T>.branch(branchName: Name): DataTree<T> =
|
|
||||||
getItem(branchName)?.tree ?: DataTree.emptyWithType(dataType)
|
|
||||||
|
|
||||||
public fun <T : Any> DataTree<T>.branch(branchName: String): DataTree<T> = branch(branchName.parseAsName())
|
|
@ -1,121 +0,0 @@
|
|||||||
package space.kscience.dataforge.data
|
|
||||||
|
|
||||||
import kotlinx.coroutines.CoroutineScope
|
|
||||||
import kotlinx.coroutines.Job
|
|
||||||
import kotlinx.coroutines.flow.MutableSharedFlow
|
|
||||||
import kotlinx.coroutines.launch
|
|
||||||
import space.kscience.dataforge.meta.Meta
|
|
||||||
import space.kscience.dataforge.misc.DFInternal
|
|
||||||
import space.kscience.dataforge.misc.ThreadSafe
|
|
||||||
import space.kscience.dataforge.names.*
|
|
||||||
import kotlin.collections.set
|
|
||||||
import kotlin.coroutines.CoroutineContext
|
|
||||||
import kotlin.reflect.KType
|
|
||||||
import kotlin.reflect.typeOf
|
|
||||||
|
|
||||||
public interface DataSourceBuilder<T : Any> : DataSetBuilder<T>, DataSource<T> {
|
|
||||||
override val updates: MutableSharedFlow<Name>
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A mutable [DataTree] that propagates updates
|
|
||||||
*/
|
|
||||||
public class DataTreeBuilder<T : Any> internal constructor(
|
|
||||||
override val dataType: KType,
|
|
||||||
coroutineContext: CoroutineContext,
|
|
||||||
) : DataTree<T>, DataSourceBuilder<T> {
|
|
||||||
|
|
||||||
override val coroutineContext: CoroutineContext =
|
|
||||||
coroutineContext + Job(coroutineContext[Job]) + GoalExecutionRestriction()
|
|
||||||
|
|
||||||
private val treeItems = HashMap<NameToken, DataTreeItem<T>>()
|
|
||||||
|
|
||||||
override val items: Map<NameToken, DataTreeItem<T>>
|
|
||||||
get() = treeItems.filter { !it.key.body.startsWith("@") }
|
|
||||||
|
|
||||||
override val updates: MutableSharedFlow<Name> = MutableSharedFlow<Name>()
|
|
||||||
|
|
||||||
@ThreadSafe
|
|
||||||
private fun remove(token: NameToken) {
|
|
||||||
if (treeItems.remove(token) != null) {
|
|
||||||
launch {
|
|
||||||
updates.emit(token.asName())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun remove(name: Name) {
|
|
||||||
if (name.isEmpty()) error("Can't remove the root node")
|
|
||||||
(getItem(name.cutLast()).tree as? DataTreeBuilder)?.remove(name.lastOrNull()!!)
|
|
||||||
}
|
|
||||||
|
|
||||||
@ThreadSafe
|
|
||||||
private fun set(token: NameToken, data: Data<T>) {
|
|
||||||
treeItems[token] = DataTreeItem.Leaf(data)
|
|
||||||
}
|
|
||||||
|
|
||||||
@ThreadSafe
|
|
||||||
private fun set(token: NameToken, node: DataTree<T>) {
|
|
||||||
treeItems[token] = DataTreeItem.Node(node)
|
|
||||||
}
|
|
||||||
|
|
||||||
private fun getOrCreateNode(token: NameToken): DataTreeBuilder<T> =
|
|
||||||
(treeItems[token] as? DataTreeItem.Node<T>)?.tree as? DataTreeBuilder<T>
|
|
||||||
?: DataTreeBuilder<T>(dataType, coroutineContext).also { set(token, it) }
|
|
||||||
|
|
||||||
private fun getOrCreateNode(name: Name): DataTreeBuilder<T> = when (name.length) {
|
|
||||||
0 -> this
|
|
||||||
1 -> getOrCreateNode(name.firstOrNull()!!)
|
|
||||||
else -> getOrCreateNode(name.firstOrNull()!!).getOrCreateNode(name.cutFirst())
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun data(name: Name, data: Data<T>?) {
|
|
||||||
if (data == null) {
|
|
||||||
remove(name)
|
|
||||||
} else {
|
|
||||||
when (name.length) {
|
|
||||||
0 -> error("Can't add data with empty name")
|
|
||||||
1 -> set(name.firstOrNull()!!, data)
|
|
||||||
2 -> getOrCreateNode(name.cutLast()).set(name.lastOrNull()!!, data)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
launch {
|
|
||||||
updates.emit(name)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun meta(name: Name, meta: Meta) {
|
|
||||||
val item = getItem(name)
|
|
||||||
if (item is DataTreeItem.Leaf) error("TODO: Can't change meta of existing leaf item.")
|
|
||||||
data(name + DataTree.META_ITEM_NAME_TOKEN, Data.empty(meta))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Create a dynamic [DataSource]. Initial data is placed synchronously.
|
|
||||||
*/
|
|
||||||
@DFInternal
|
|
||||||
@Suppress("FunctionName")
|
|
||||||
public fun <T : Any> DataSource(
|
|
||||||
type: KType,
|
|
||||||
parent: CoroutineScope,
|
|
||||||
block: DataSourceBuilder<T>.() -> Unit = {},
|
|
||||||
): DataTreeBuilder<T> = DataTreeBuilder<T>(type, parent.coroutineContext).apply(block)
|
|
||||||
|
|
||||||
@Suppress("OPT_IN_USAGE", "FunctionName")
|
|
||||||
public inline fun <reified T : Any> DataSource(
|
|
||||||
parent: CoroutineScope,
|
|
||||||
crossinline block: DataSourceBuilder<T>.() -> Unit = {},
|
|
||||||
): DataTreeBuilder<T> = DataSource(typeOf<T>(), parent) { block() }
|
|
||||||
|
|
||||||
public inline fun <reified T : Any> DataSourceBuilder<T>.emit(
|
|
||||||
name: Name,
|
|
||||||
parent: CoroutineScope,
|
|
||||||
noinline block: DataSourceBuilder<T>.() -> Unit,
|
|
||||||
): Unit = node(name, DataSource(parent, block))
|
|
||||||
|
|
||||||
public inline fun <reified T : Any> DataSourceBuilder<T>.emit(
|
|
||||||
name: String,
|
|
||||||
parent: CoroutineScope,
|
|
||||||
noinline block: DataSourceBuilder<T>.() -> Unit,
|
|
||||||
): Unit = node(Name.parse(name), DataSource(parent, block))
|
|
@ -9,7 +9,7 @@ import kotlin.coroutines.EmptyCoroutineContext
|
|||||||
* Lazy computation result with its dependencies to allowing to stat computing dependencies ahead of time
|
* Lazy computation result with its dependencies to allowing to stat computing dependencies ahead of time
|
||||||
*/
|
*/
|
||||||
public interface Goal<out T> {
|
public interface Goal<out T> {
|
||||||
public val dependencies: Collection<Goal<*>>
|
public val dependencies: Iterable<Goal<*>>
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Returns current running coroutine if the goal is started. Null if the computation is not started.
|
* Returns current running coroutine if the goal is started. Null if the computation is not started.
|
||||||
@ -54,7 +54,7 @@ public open class StaticGoal<T>(public val value: T) : Goal<T> {
|
|||||||
*/
|
*/
|
||||||
public open class LazyGoal<T>(
|
public open class LazyGoal<T>(
|
||||||
private val coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
private val coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
||||||
override val dependencies: Collection<Goal<*>> = emptyList(),
|
override val dependencies: Iterable<Goal<*>> = emptyList(),
|
||||||
public val block: suspend () -> T,
|
public val block: suspend () -> T,
|
||||||
) : Goal<T> {
|
) : Goal<T> {
|
||||||
|
|
||||||
@ -82,8 +82,8 @@ public open class LazyGoal<T>(
|
|||||||
}
|
}
|
||||||
|
|
||||||
log?.emit { "Starting dependencies computation for ${this@LazyGoal}" }
|
log?.emit { "Starting dependencies computation for ${this@LazyGoal}" }
|
||||||
val startedDependencies = this.dependencies.map { goal ->
|
val startedDependencies = dependencies.map { goal ->
|
||||||
goal.run { async(coroutineScope) }
|
goal.async(coroutineScope)
|
||||||
}
|
}
|
||||||
return deferred ?: coroutineScope.async(
|
return deferred ?: coroutineScope.async(
|
||||||
coroutineContext
|
coroutineContext
|
||||||
|
@ -15,13 +15,11 @@
|
|||||||
*/
|
*/
|
||||||
package space.kscience.dataforge.data
|
package space.kscience.dataforge.data
|
||||||
|
|
||||||
import kotlinx.coroutines.launch
|
|
||||||
import space.kscience.dataforge.meta.get
|
import space.kscience.dataforge.meta.get
|
||||||
import space.kscience.dataforge.meta.string
|
import space.kscience.dataforge.meta.string
|
||||||
import space.kscience.dataforge.misc.DFInternal
|
|
||||||
|
|
||||||
public interface GroupRule {
|
public interface GroupRule {
|
||||||
public fun <T : Any> gather(set: DataSet<T>): Map<String, DataSet<T>>
|
public fun <T : Any> gather(set: DataTree<T>): Map<String, DataTree<T>>
|
||||||
|
|
||||||
public companion object {
|
public companion object {
|
||||||
/**
|
/**
|
||||||
@ -32,46 +30,23 @@ public interface GroupRule {
|
|||||||
* @param defaultTagValue
|
* @param defaultTagValue
|
||||||
* @return
|
* @return
|
||||||
*/
|
*/
|
||||||
@OptIn(DFInternal::class)
|
|
||||||
public fun byMetaValue(
|
public fun byMetaValue(
|
||||||
key: String,
|
key: String,
|
||||||
defaultTagValue: String,
|
defaultTagValue: String,
|
||||||
): GroupRule = object : GroupRule {
|
): GroupRule = object : GroupRule {
|
||||||
|
|
||||||
override fun <T : Any> gather(
|
override fun <T : Any> gather(
|
||||||
set: DataSet<T>,
|
set: DataTree<T>,
|
||||||
): Map<String, DataSet<T>> {
|
): Map<String, DataTree<T>> {
|
||||||
val map = HashMap<String, DataSet<T>>()
|
val map = HashMap<String, DataTreeBuilder<T>>()
|
||||||
|
|
||||||
if (set is DataSource) {
|
|
||||||
set.forEach { data ->
|
set.forEach { data ->
|
||||||
val tagValue: String = data.meta[key]?.string ?: defaultTagValue
|
val tagValue: String = data.meta[key]?.string ?: defaultTagValue
|
||||||
(map.getOrPut(tagValue) { DataTreeBuilder(set.dataType, set.coroutineContext) } as DataTreeBuilder<T>)
|
map.getOrPut(tagValue) { DataTreeBuilder(set.dataType) }.emit(data.name,data.data)
|
||||||
.data(data.name, data.data)
|
|
||||||
|
|
||||||
set.launch {
|
|
||||||
set.updates.collect { name ->
|
|
||||||
val dataUpdate = set[name]
|
|
||||||
|
|
||||||
val updateTagValue = dataUpdate?.meta?.get(key)?.string ?: defaultTagValue
|
|
||||||
map.getOrPut(updateTagValue) {
|
|
||||||
DataSource(set.dataType, this) {
|
|
||||||
data(name, dataUpdate)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
set.forEach { data ->
|
|
||||||
val tagValue: String = data.meta[key]?.string ?: defaultTagValue
|
|
||||||
(map.getOrPut(tagValue) { StaticDataTree(set.dataType) } as StaticDataTree<T>)
|
|
||||||
.data(data.name, data.data)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
return map
|
return map.mapValues { it.value.build() }
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -4,7 +4,7 @@ import space.kscience.dataforge.meta.isEmpty
|
|||||||
import space.kscience.dataforge.misc.Named
|
import space.kscience.dataforge.misc.Named
|
||||||
import space.kscience.dataforge.names.Name
|
import space.kscience.dataforge.names.Name
|
||||||
|
|
||||||
public interface NamedData<out T : Any> : Named, Data<T> {
|
public interface NamedData<T> : Named, Data<T> {
|
||||||
override val name: Name
|
override val name: Name
|
||||||
public val data: Data<T>
|
public val data: Data<T>
|
||||||
}
|
}
|
||||||
@ -12,7 +12,7 @@ public interface NamedData<out T : Any> : Named, Data<T> {
|
|||||||
public operator fun NamedData<*>.component1(): Name = name
|
public operator fun NamedData<*>.component1(): Name = name
|
||||||
public operator fun <T: Any> NamedData<T>.component2(): Data<T> = data
|
public operator fun <T: Any> NamedData<T>.component2(): Data<T> = data
|
||||||
|
|
||||||
private class NamedDataImpl<out T : Any>(
|
private class NamedDataImpl<T>(
|
||||||
override val name: Name,
|
override val name: Name,
|
||||||
override val data: Data<T>,
|
override val data: Data<T>,
|
||||||
) : Data<T> by data, NamedData<T> {
|
) : Data<T> by data, NamedData<T> {
|
||||||
@ -28,7 +28,7 @@ private class NamedDataImpl<out T : Any>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public fun <T : Any> Data<T>.named(name: Name): NamedData<T> = if (this is NamedData) {
|
public fun <T> Data<T>.named(name: Name): NamedData<T> = if (this is NamedData) {
|
||||||
NamedDataImpl(name, this.data)
|
NamedDataImpl(name, this.data)
|
||||||
} else {
|
} else {
|
||||||
NamedDataImpl(name, this)
|
NamedDataImpl(name, this)
|
||||||
|
@ -1,80 +0,0 @@
|
|||||||
package space.kscience.dataforge.data
|
|
||||||
|
|
||||||
import space.kscience.dataforge.meta.Meta
|
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
|
||||||
import space.kscience.dataforge.names.*
|
|
||||||
import kotlin.reflect.KType
|
|
||||||
import kotlin.reflect.typeOf
|
|
||||||
|
|
||||||
@PublishedApi
|
|
||||||
internal class StaticDataTree<T : Any>(
|
|
||||||
override val dataType: KType,
|
|
||||||
) : DataSetBuilder<T>, DataTree<T> {
|
|
||||||
|
|
||||||
private val _items: MutableMap<NameToken, DataTreeItem<T>> = HashMap()
|
|
||||||
|
|
||||||
override val items: Map<NameToken, DataTreeItem<T>>
|
|
||||||
get() = _items.filter { !it.key.body.startsWith("@") }
|
|
||||||
|
|
||||||
override fun remove(name: Name) {
|
|
||||||
when (name.length) {
|
|
||||||
0 -> error("Can't remove root tree node")
|
|
||||||
1 -> _items.remove(name.firstOrNull()!!)
|
|
||||||
else -> (_items[name.firstOrNull()!!].tree as? StaticDataTree<T>)?.remove(name.cutFirst())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private fun getOrCreateNode(name: Name): StaticDataTree<T> = when (name.length) {
|
|
||||||
0 -> this
|
|
||||||
1 -> {
|
|
||||||
val itemName = name.firstOrNull()!!
|
|
||||||
(_items[itemName].tree as? StaticDataTree<T>) ?: StaticDataTree<T>(dataType).also {
|
|
||||||
_items[itemName] = DataTreeItem.Node(it)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
else -> getOrCreateNode(name.cutLast()).getOrCreateNode(name.lastOrNull()!!.asName())
|
|
||||||
}
|
|
||||||
|
|
||||||
private fun set(name: Name, item: DataTreeItem<T>?) {
|
|
||||||
if (name.isEmpty()) error("Can't set top level tree node")
|
|
||||||
if (item == null) {
|
|
||||||
remove(name)
|
|
||||||
} else {
|
|
||||||
getOrCreateNode(name.cutLast())._items[name.lastOrNull()!!] = item
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun data(name: Name, data: Data<T>?) {
|
|
||||||
set(name, data?.let { DataTreeItem.Leaf(it) })
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun node(name: Name, dataSet: DataSet<T>) {
|
|
||||||
if (dataSet is StaticDataTree) {
|
|
||||||
set(name, DataTreeItem.Node(dataSet))
|
|
||||||
} else {
|
|
||||||
dataSet.forEach {
|
|
||||||
data(name + it.name, it.data)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun meta(name: Name, meta: Meta) {
|
|
||||||
val item = getItem(name)
|
|
||||||
if (item is DataTreeItem.Leaf) TODO("Can't change meta of existing leaf item.")
|
|
||||||
data(name + DataTree.META_ITEM_NAME_TOKEN, Data.empty(meta))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public inline fun <T : Any> DataTree(
|
|
||||||
dataType: KType,
|
|
||||||
block: DataSetBuilder<T>.() -> Unit,
|
|
||||||
): DataTree<T> = StaticDataTree<T>(dataType).apply { block() }
|
|
||||||
|
|
||||||
public inline fun <reified T : Any> DataTree(
|
|
||||||
noinline block: DataSetBuilder<T>.() -> Unit,
|
|
||||||
): DataTree<T> = DataTree(typeOf<T>(), block)
|
|
||||||
|
|
||||||
@OptIn(DFExperimental::class)
|
|
||||||
public fun <T : Any> DataSet<T>.seal(): DataTree<T> = DataTree(dataType) {
|
|
||||||
populateFrom(this@seal)
|
|
||||||
}
|
|
@ -0,0 +1,121 @@
|
|||||||
|
package space.kscience.dataforge.data
|
||||||
|
|
||||||
|
import kotlinx.coroutines.Job
|
||||||
|
import kotlinx.coroutines.flow.launchIn
|
||||||
|
import kotlinx.coroutines.flow.onEach
|
||||||
|
import space.kscience.dataforge.meta.Meta
|
||||||
|
import space.kscience.dataforge.meta.MutableMeta
|
||||||
|
import space.kscience.dataforge.misc.DFExperimental
|
||||||
|
import space.kscience.dataforge.names.Name
|
||||||
|
import space.kscience.dataforge.names.isEmpty
|
||||||
|
import space.kscience.dataforge.names.plus
|
||||||
|
|
||||||
|
|
||||||
|
public fun <T> DataSink<T>.emit(value: NamedData<T>) {
|
||||||
|
emit(value.name, value.data)
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun <T> DataSink<T>.emitAll(sequence: Sequence<NamedData<T>>) {
|
||||||
|
sequence.forEach { emit(it) }
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun <T> DataSink<T>.emitAll(dataTree: DataTree<T>) {
|
||||||
|
emitAll(dataTree.asSequence())
|
||||||
|
}
|
||||||
|
|
||||||
|
public inline fun <T> DataSink<T>.emitAll(
|
||||||
|
prefix: Name,
|
||||||
|
block: DataSink<T>.() -> Unit,
|
||||||
|
) {
|
||||||
|
if (prefix.isEmpty()) {
|
||||||
|
apply(block)
|
||||||
|
} else {
|
||||||
|
val proxyDataSink = DataSink { nameWithoutPrefix, data ->
|
||||||
|
this.emit(prefix + nameWithoutPrefix, data)
|
||||||
|
}
|
||||||
|
|
||||||
|
proxyDataSink.apply(block)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
public fun <T> DataSink<T>.emit(name: String, value: Data<T>) {
|
||||||
|
emit(Name.parse(name), value)
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun <T> DataSink<T>.emitAll(name: Name, set: DataTree<T>) {
|
||||||
|
emitAll(name) { emitAll(set.asSequence()) }
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun <T> DataSink<T>.emitAll(name: String, set: DataTree<T>) {
|
||||||
|
emitAll(Name.parse(name)) { emitAll(set.asSequence()) }
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Produce lazy [Data] and emit it into the [MutableDataTree]
|
||||||
|
*/
|
||||||
|
public inline fun <reified T> DataSink<T>.produce(
|
||||||
|
name: String,
|
||||||
|
meta: Meta = Meta.EMPTY,
|
||||||
|
noinline producer: suspend () -> T,
|
||||||
|
) {
|
||||||
|
val data = Data(meta, block = producer)
|
||||||
|
emit(name, data)
|
||||||
|
}
|
||||||
|
|
||||||
|
public inline fun <reified T> DataSink<T>.produce(
|
||||||
|
name: Name,
|
||||||
|
meta: Meta = Meta.EMPTY,
|
||||||
|
noinline producer: suspend () -> T,
|
||||||
|
) {
|
||||||
|
val data = Data(meta, block = producer)
|
||||||
|
emit(name, data)
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Emit static data with the fixed value
|
||||||
|
*/
|
||||||
|
public inline fun <reified T> DataSink<T>.static(
|
||||||
|
name: String,
|
||||||
|
data: T,
|
||||||
|
meta: Meta = Meta.EMPTY,
|
||||||
|
): Unit = emit(name, Data.static(data, meta))
|
||||||
|
|
||||||
|
public inline fun <reified T> DataSink<T>.static(
|
||||||
|
name: Name,
|
||||||
|
data: T,
|
||||||
|
meta: Meta = Meta.EMPTY,
|
||||||
|
): Unit = emit(name, Data.static(data, meta))
|
||||||
|
|
||||||
|
public inline fun <reified T> DataSink<T>.static(
|
||||||
|
name: String,
|
||||||
|
data: T,
|
||||||
|
mutableMeta: MutableMeta.() -> Unit,
|
||||||
|
): Unit = emit(Name.parse(name), Data.static(data, Meta(mutableMeta)))
|
||||||
|
|
||||||
|
|
||||||
|
public fun <T> DataSink<T>.populateFrom(sequence: Sequence<NamedData<T>>) {
|
||||||
|
sequence.forEach {
|
||||||
|
emit(it.name, it.data)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public fun <T> DataSink<T>.populateFrom(tree: DataTree<T>) {
|
||||||
|
populateFrom(tree.asSequence())
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Update data with given node data and meta with node meta.
|
||||||
|
*/
|
||||||
|
@DFExperimental
|
||||||
|
public fun <T> MutableDataTree<T>.populateFrom(flow: ObservableDataSource<T>): Job = flow.updates().onEach {
|
||||||
|
//TODO check if the place is occupied
|
||||||
|
emit(it.name, it.data)
|
||||||
|
}.launchIn(scope)
|
||||||
|
|
||||||
|
//public fun <T > DataSetBuilder<T>.populateFrom(flow: Flow<NamedData<T>>) {
|
||||||
|
// flow.collect {
|
||||||
|
// data(it.name, it.data)
|
||||||
|
// }
|
||||||
|
//}
|
@ -1,105 +0,0 @@
|
|||||||
package space.kscience.dataforge.data
|
|
||||||
|
|
||||||
import kotlinx.coroutines.flow.Flow
|
|
||||||
import kotlinx.coroutines.flow.filter
|
|
||||||
import kotlinx.coroutines.flow.map
|
|
||||||
import kotlinx.coroutines.flow.mapNotNull
|
|
||||||
import space.kscience.dataforge.meta.Meta
|
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
|
||||||
import space.kscience.dataforge.names.*
|
|
||||||
import kotlin.coroutines.CoroutineContext
|
|
||||||
import kotlin.coroutines.EmptyCoroutineContext
|
|
||||||
import kotlin.reflect.KType
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A stateless filtered [DataSet]
|
|
||||||
*/
|
|
||||||
public fun <T : Any> DataSet<T>.filter(
|
|
||||||
predicate: (Name, Meta) -> Boolean,
|
|
||||||
): DataSource<T> = object : DataSource<T> {
|
|
||||||
|
|
||||||
override val dataType: KType get() = this@filter.dataType
|
|
||||||
|
|
||||||
override val coroutineContext: CoroutineContext
|
|
||||||
get() = (this@filter as? DataSource)?.coroutineContext ?: EmptyCoroutineContext
|
|
||||||
|
|
||||||
|
|
||||||
override val meta: Meta get() = this@filter.meta
|
|
||||||
|
|
||||||
override fun iterator(): Iterator<NamedData<T>> = iterator {
|
|
||||||
for (d in this@filter) {
|
|
||||||
if (predicate(d.name, d.meta)) {
|
|
||||||
yield(d)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun get(name: Name): Data<T>? = this@filter.get(name)?.takeIf {
|
|
||||||
predicate(name, it.meta)
|
|
||||||
}
|
|
||||||
|
|
||||||
override val updates: Flow<Name> = this@filter.updates.filter flowFilter@{ name ->
|
|
||||||
val theData = this@filter[name] ?: return@flowFilter false
|
|
||||||
predicate(name, theData.meta)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Generate a wrapper data set with a given name prefix appended to all names
|
|
||||||
*/
|
|
||||||
public fun <T : Any> DataSet<T>.withNamePrefix(prefix: Name): DataSet<T> = if (prefix.isEmpty()) {
|
|
||||||
this
|
|
||||||
} else object : DataSource<T> {
|
|
||||||
|
|
||||||
override val dataType: KType get() = this@withNamePrefix.dataType
|
|
||||||
|
|
||||||
override val coroutineContext: CoroutineContext
|
|
||||||
get() = (this@withNamePrefix as? DataSource)?.coroutineContext ?: EmptyCoroutineContext
|
|
||||||
|
|
||||||
override val meta: Meta get() = this@withNamePrefix.meta
|
|
||||||
|
|
||||||
|
|
||||||
override fun iterator(): Iterator<NamedData<T>> = iterator {
|
|
||||||
for (d in this@withNamePrefix) {
|
|
||||||
yield(d.data.named(prefix + d.name))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun get(name: Name): Data<T>? =
|
|
||||||
name.removeFirstOrNull(name)?.let { this@withNamePrefix.get(it) }
|
|
||||||
|
|
||||||
override val updates: Flow<Name> get() = this@withNamePrefix.updates.map { prefix + it }
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Get a subset of data starting with a given [branchName]
|
|
||||||
*/
|
|
||||||
public fun <T : Any> DataSet<T>.branch(branchName: Name): DataSet<T> = if (branchName.isEmpty()) {
|
|
||||||
this
|
|
||||||
} else object : DataSource<T> {
|
|
||||||
override val dataType: KType get() = this@branch.dataType
|
|
||||||
|
|
||||||
override val coroutineContext: CoroutineContext
|
|
||||||
get() = (this@branch as? DataSource)?.coroutineContext ?: EmptyCoroutineContext
|
|
||||||
|
|
||||||
override val meta: Meta get() = this@branch.meta
|
|
||||||
|
|
||||||
override fun iterator(): Iterator<NamedData<T>> = iterator {
|
|
||||||
for (d in this@branch) {
|
|
||||||
d.name.removeFirstOrNull(branchName)?.let { name ->
|
|
||||||
yield(d.data.named(name))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun get(name: Name): Data<T>? = this@branch.get(branchName + name)
|
|
||||||
|
|
||||||
override val updates: Flow<Name> get() = this@branch.updates.mapNotNull { it.removeFirstOrNull(branchName) }
|
|
||||||
}
|
|
||||||
|
|
||||||
public fun <T : Any> DataSet<T>.branch(branchName: String): DataSet<T> = this@branch.branch(branchName.parseAsName())
|
|
||||||
|
|
||||||
@DFExperimental
|
|
||||||
public suspend fun <T : Any> DataSet<T>.rootData(): Data<T>? = get(Name.EMPTY)
|
|
||||||
|
|
@ -11,11 +11,11 @@ import kotlin.reflect.typeOf
|
|||||||
|
|
||||||
public data class ValueWithMeta<T>(val meta: Meta, val value: T)
|
public data class ValueWithMeta<T>(val meta: Meta, val value: T)
|
||||||
|
|
||||||
public suspend fun <T : Any> Data<T>.awaitWithMeta(): ValueWithMeta<T> = ValueWithMeta(meta, await())
|
public suspend fun <T> Data<T>.awaitWithMeta(): ValueWithMeta<T> = ValueWithMeta(meta, await())
|
||||||
|
|
||||||
public data class NamedValueWithMeta<T>(val name: Name, val meta: Meta, val value: T)
|
public data class NamedValueWithMeta<T>(val name: Name, val meta: Meta, val value: T)
|
||||||
|
|
||||||
public suspend fun <T : Any> NamedData<T>.awaitWithMeta(): NamedValueWithMeta<T> =
|
public suspend fun <T> NamedData<T>.awaitWithMeta(): NamedValueWithMeta<T> =
|
||||||
NamedValueWithMeta(name, meta, await())
|
NamedValueWithMeta(name, meta, await())
|
||||||
|
|
||||||
|
|
||||||
@ -25,7 +25,7 @@ public suspend fun <T : Any> NamedData<T>.awaitWithMeta(): NamedValueWithMeta<T>
|
|||||||
* @param meta for the resulting data. By default equals input data.
|
* @param meta for the resulting data. By default equals input data.
|
||||||
* @param block the transformation itself
|
* @param block the transformation itself
|
||||||
*/
|
*/
|
||||||
public inline fun <T : Any, reified R : Any> Data<T>.map(
|
public inline fun <T : Any, reified R : Any> Data<T>.transform(
|
||||||
meta: Meta = this.meta,
|
meta: Meta = this.meta,
|
||||||
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
||||||
crossinline block: suspend (T) -> R,
|
crossinline block: suspend (T) -> R,
|
||||||
@ -58,20 +58,7 @@ internal fun Iterable<Data<*>>.joinMeta(): Meta = Meta {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Lazily reduce a collection of [Data] to a single data.
|
|
||||||
*/
|
|
||||||
public inline fun <T : Any, reified R : Any> Collection<Data<T>>.reduceToData(
|
|
||||||
meta: Meta = joinMeta(),
|
|
||||||
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
|
||||||
crossinline block: suspend (List<ValueWithMeta<T>>) -> R,
|
|
||||||
): Data<R> = Data(
|
|
||||||
meta,
|
|
||||||
coroutineContext,
|
|
||||||
this
|
|
||||||
) {
|
|
||||||
block(map { it.awaitWithMeta() })
|
|
||||||
}
|
|
||||||
|
|
||||||
@PublishedApi
|
@PublishedApi
|
||||||
internal fun Map<*, Data<*>>.joinMeta(): Meta = Meta {
|
internal fun Map<*, Data<*>>.joinMeta(): Meta = Meta {
|
||||||
@ -195,53 +182,52 @@ public inline fun <T : Any, reified R : Any> Iterable<NamedData<T>>.foldNamedToD
|
|||||||
//DataSet operations
|
//DataSet operations
|
||||||
|
|
||||||
@DFInternal
|
@DFInternal
|
||||||
public suspend fun <T : Any, R : Any> DataSet<T>.map(
|
public suspend fun <T, R> DataTree<T>.transform(
|
||||||
outputType: KType,
|
outputType: KType,
|
||||||
metaTransform: MutableMeta.() -> Unit = {},
|
metaTransform: MutableMeta.() -> Unit = {},
|
||||||
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
||||||
block: suspend (NamedValueWithMeta<T>) -> R,
|
block: suspend (NamedValueWithMeta<T>) -> R,
|
||||||
): DataTree<R> = DataTree<R>(outputType){
|
): DataTree<R> = DataTree<R>(outputType){
|
||||||
forEach {
|
//quasi-synchronous processing of elements in the tree
|
||||||
val newMeta = it.meta.toMutableMeta().apply(metaTransform).seal()
|
asSequence().forEach { namedData: NamedData<T> ->
|
||||||
val d = Data(outputType, newMeta, coroutineContext, listOf(it)) {
|
val newMeta = namedData.meta.toMutableMeta().apply(metaTransform).seal()
|
||||||
block(it.awaitWithMeta())
|
val d = Data(outputType, newMeta, coroutineContext, listOf(namedData)) {
|
||||||
|
block(namedData.awaitWithMeta())
|
||||||
}
|
}
|
||||||
data(it.name, d)
|
emit(namedData.name, d)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@OptIn(DFInternal::class)
|
@OptIn(DFInternal::class)
|
||||||
public suspend inline fun <T : Any, reified R : Any> DataSet<T>.map(
|
public suspend inline fun <T : Any, reified R : Any> DataTree<T>.transform(
|
||||||
noinline metaTransform: MutableMeta.() -> Unit = {},
|
noinline metaTransform: MutableMeta.() -> Unit = {},
|
||||||
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
||||||
noinline block: suspend (NamedValueWithMeta<T>) -> R,
|
noinline block: suspend (NamedValueWithMeta<T>) -> R,
|
||||||
): DataTree<R> = map(typeOf<R>(), metaTransform, coroutineContext, block)
|
): DataTree<R> = this@transform.transform(typeOf<R>(), metaTransform, coroutineContext, block)
|
||||||
|
|
||||||
public inline fun <T : Any> DataSet<T>.forEach(block: (NamedData<T>) -> Unit) {
|
public inline fun <T : Any> DataTree<T>.forEach(block: (NamedData<T>) -> Unit) {
|
||||||
for (d in this) {
|
asSequence().forEach(block)
|
||||||
block(d)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// DataSet reduction
|
// DataSet reduction
|
||||||
|
|
||||||
@PublishedApi
|
@PublishedApi
|
||||||
internal fun DataSet<*>.joinMeta(): Meta = Meta {
|
internal fun DataTree<*>.joinMeta(): Meta = Meta {
|
||||||
forEach { (key, data) ->
|
asSequence().forEach {
|
||||||
val token = NameToken("data", key.toString())
|
val token = NameToken("data", it.name.toString())
|
||||||
set(token, data.meta)
|
set(token, it.meta)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public inline fun <T : Any, reified R : Any> DataSet<T>.reduceToData(
|
public inline fun <T : Any, reified R : Any> DataTree<T>.reduceToData(
|
||||||
meta: Meta = joinMeta(),
|
meta: Meta = joinMeta(),
|
||||||
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
||||||
crossinline transformation: suspend (Iterable<NamedValueWithMeta<T>>) -> R,
|
crossinline transformation: suspend (Iterable<NamedValueWithMeta<T>>) -> R,
|
||||||
): Data<R> = asIterable().reduceNamedToData(meta, coroutineContext, transformation)
|
): Data<R> = asSequence().asIterable().reduceNamedToData(meta, coroutineContext, transformation)
|
||||||
|
|
||||||
public inline fun <T : Any, reified R : Any> DataSet<T>.foldToData(
|
public inline fun <T : Any, reified R : Any> DataTree<T>.foldToData(
|
||||||
initial: R,
|
initial: R,
|
||||||
meta: Meta = joinMeta(),
|
meta: Meta = joinMeta(),
|
||||||
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
coroutineContext: CoroutineContext = EmptyCoroutineContext,
|
||||||
crossinline block: suspend (result: R, data: NamedValueWithMeta<T>) -> R,
|
crossinline block: suspend (result: R, data: NamedValueWithMeta<T>) -> R,
|
||||||
): Data<R> = asIterable().foldNamedToData(initial, meta, coroutineContext, block)
|
): Data<R> = asSequence().asIterable().foldNamedToData(initial, meta, coroutineContext, block)
|
@ -1,12 +1,10 @@
|
|||||||
package space.kscience.dataforge.data
|
package space.kscience.dataforge.data
|
||||||
|
|
||||||
|
import kotlinx.coroutines.CoroutineScope
|
||||||
import kotlinx.coroutines.flow.Flow
|
import kotlinx.coroutines.flow.Flow
|
||||||
import kotlinx.coroutines.flow.filter
|
import kotlinx.coroutines.flow.filter
|
||||||
import space.kscience.dataforge.meta.Meta
|
import space.kscience.dataforge.misc.DFInternal
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
|
||||||
import space.kscience.dataforge.names.Name
|
import space.kscience.dataforge.names.Name
|
||||||
import kotlin.coroutines.CoroutineContext
|
|
||||||
import kotlin.coroutines.EmptyCoroutineContext
|
|
||||||
import kotlin.reflect.KType
|
import kotlin.reflect.KType
|
||||||
import kotlin.reflect.full.isSubtypeOf
|
import kotlin.reflect.full.isSubtypeOf
|
||||||
import kotlin.reflect.typeOf
|
import kotlin.reflect.typeOf
|
||||||
@ -25,61 +23,65 @@ private fun <R : Any> Data<*>.castOrNull(type: KType): Data<R>? =
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Suppress("UNCHECKED_CAST")
|
||||||
|
@DFInternal
|
||||||
|
public fun <R> Sequence<NamedData<*>>.filterByDataType(type: KType): Sequence<NamedData<R>> =
|
||||||
|
filter { it.type.isSubtypeOf(type) } as Sequence<NamedData<R>>
|
||||||
|
|
||||||
|
@Suppress("UNCHECKED_CAST")
|
||||||
|
@DFInternal
|
||||||
|
public fun <R> Flow<NamedData<*>>.filterByDataType(type: KType): Flow<NamedData<R>> =
|
||||||
|
filter { it.type.isSubtypeOf(type) } as Flow<NamedData<R>>
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Select all data matching given type and filters. Does not modify paths
|
* Select all data matching given type and filters. Does not modify paths
|
||||||
*
|
*
|
||||||
* @param predicate addition filtering condition based on item name and meta. By default, accepts all
|
* @param predicate additional filtering condition based on item name and meta. By default, accepts all
|
||||||
*/
|
*/
|
||||||
@OptIn(DFExperimental::class)
|
@DFInternal
|
||||||
public fun <R : Any> DataSet<*>.filterByType(
|
public fun <R> DataTree<*>.filterByType(
|
||||||
type: KType,
|
type: KType,
|
||||||
predicate: (name: Name, meta: Meta) -> Boolean = { _, _ -> true },
|
predicate: DataFilter = DataFilter.EMPTY,
|
||||||
): DataSource<R> = object : DataSource<R> {
|
): DataTree<R> = asSequence().filterByDataType<R>(type).filterData(predicate).toTree(type)
|
||||||
override val dataType = type
|
|
||||||
|
|
||||||
override val coroutineContext: CoroutineContext
|
|
||||||
get() = (this@filterByType as? DataSource)?.coroutineContext ?: EmptyCoroutineContext
|
|
||||||
|
|
||||||
override val meta: Meta get() = this@filterByType.meta
|
|
||||||
|
|
||||||
private fun checkDatum(name: Name, datum: Data<*>): Boolean = datum.type.isSubtypeOf(type)
|
|
||||||
&& predicate(name, datum.meta)
|
|
||||||
|
|
||||||
override fun iterator(): Iterator<NamedData<R>> = iterator {
|
|
||||||
for(d in this@filterByType){
|
|
||||||
if(checkDatum(d.name,d.data)){
|
|
||||||
@Suppress("UNCHECKED_CAST")
|
|
||||||
yield(d as NamedData<R>)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun get(name: Name): Data<R>? = this@filterByType[name]?.let { datum ->
|
|
||||||
if (checkDatum(name, datum)) datum.castOrNull(type) else null
|
|
||||||
}
|
|
||||||
|
|
||||||
override val updates: Flow<Name> = this@filterByType.updates.filter { name ->
|
|
||||||
get(name)?.let { datum ->
|
|
||||||
checkDatum(name, datum)
|
|
||||||
} ?: false
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Select a single datum of the appropriate type
|
* Select a single datum of the appropriate type
|
||||||
*/
|
*/
|
||||||
public inline fun <reified R : Any> DataSet<*>.filterByType(
|
@OptIn(DFInternal::class)
|
||||||
noinline predicate: (name: Name, meta: Meta) -> Boolean = { _, _ -> true },
|
public inline fun <reified R : Any> DataTree<*>.filterByType(
|
||||||
): DataSet<R> = filterByType(typeOf<R>(), predicate)
|
predicate: DataFilter = DataFilter.EMPTY,
|
||||||
|
): DataTree<R> = filterByType(typeOf<R>(), predicate)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Select a single datum if it is present and of given [type]
|
* Select a single datum if it is present and of given [type]
|
||||||
*/
|
*/
|
||||||
public fun <R : Any> DataSet<*>.getByType(type: KType, name: Name): NamedData<R>? =
|
public fun <R : Any> DataTree<*>.getByType(type: KType, name: Name): NamedData<R>? =
|
||||||
get(name)?.castOrNull<R>(type)?.named(name)
|
get(name)?.castOrNull<R>(type)?.named(name)
|
||||||
|
|
||||||
public inline fun <reified R : Any> DataSet<*>.getByType(name: Name): NamedData<R>? =
|
public inline fun <reified R : Any> DataTree<*>.getByType(name: Name): NamedData<R>? =
|
||||||
this@getByType.getByType(typeOf<R>(), name)
|
this@getByType.getByType(typeOf<R>(), name)
|
||||||
|
|
||||||
public inline fun <reified R : Any> DataSet<*>.getByType(name: String): NamedData<R>? =
|
public inline fun <reified R : Any> DataTree<*>.getByType(name: String): NamedData<R>? =
|
||||||
this@getByType.getByType(typeOf<R>(), Name.parse(name))
|
this@getByType.getByType(typeOf<R>(), Name.parse(name))
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Select all data matching given type and filters. Does not modify paths
|
||||||
|
*
|
||||||
|
* @param predicate additional filtering condition based on item name and meta. By default, accepts all
|
||||||
|
*/
|
||||||
|
@DFInternal
|
||||||
|
public fun <R> ObservableDataTree<*>.filterByType(
|
||||||
|
type: KType,
|
||||||
|
scope: CoroutineScope,
|
||||||
|
predicate: DataFilter = DataFilter.EMPTY,
|
||||||
|
): ObservableDataTree<R> = asSequence()
|
||||||
|
.filterByDataType<R>(type)
|
||||||
|
.filterData(predicate)
|
||||||
|
.toObservableTree(type, scope, updates().filterByDataType<R>(type).filterData(predicate))
|
||||||
|
|
||||||
|
|
||||||
|
@OptIn(DFInternal::class)
|
||||||
|
public inline fun <reified R> ObservableDataTree<*>.filterByType(
|
||||||
|
scope: CoroutineScope,
|
||||||
|
predicate: DataFilter = DataFilter.EMPTY,
|
||||||
|
): ObservableDataTree<R> = filterByType(typeOf<R>(),scope,predicate)
|
@ -10,31 +10,35 @@ import space.kscience.dataforge.names.plus
|
|||||||
/**
|
/**
|
||||||
* Append data to node
|
* Append data to node
|
||||||
*/
|
*/
|
||||||
context(DataSetBuilder<T>) public infix fun <T : Any> String.put(data: Data<T>): Unit =
|
context(DataSink<T>)
|
||||||
data(Name.parse(this), data)
|
public infix fun <T : Any> String.put(data: Data<T>): Unit =
|
||||||
|
emit(Name.parse(this), data)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Append node
|
* Append node
|
||||||
*/
|
*/
|
||||||
context(DataSetBuilder<T>) public infix fun <T : Any> String.put(dataSet: DataSet<T>): Unit =
|
context(DataSink<T>)
|
||||||
node(Name.parse(this), dataSet)
|
public infix fun <T : Any> String.put(dataSet: DataTree<T>): Unit =
|
||||||
|
emitAll(this, dataSet)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Build and append node
|
* Build and append node
|
||||||
*/
|
*/
|
||||||
context(DataSetBuilder<T>) public infix fun <T : Any> String.put(
|
context(DataSink<T>)
|
||||||
block: DataSetBuilder<T>.() -> Unit,
|
public infix fun <T : Any> String.put(
|
||||||
): Unit = node(Name.parse(this), block)
|
block: DataSink<T>.() -> Unit,
|
||||||
|
): Unit = emitAll(Name.parse(this), block)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Copy given data set and mirror its changes to this [DataTreeBuilder] in [this@setAndObserve]. Returns an update [Job]
|
* Copy given data set and mirror its changes to this [LegacyDataTreeBuilder] in [this@setAndObserve]. Returns an update [Job]
|
||||||
*/
|
*/
|
||||||
context(DataSetBuilder<T>) public fun <T : Any> CoroutineScope.setAndWatch(
|
context(DataSink<T>)
|
||||||
|
public fun <T : Any> CoroutineScope.setAndWatch(
|
||||||
name: Name,
|
name: Name,
|
||||||
dataSet: DataSet<T>,
|
dataSet: DataTree<T>,
|
||||||
): Job = launch {
|
): Job = launch {
|
||||||
node(name, dataSet)
|
emitAll(name, dataSet)
|
||||||
dataSet.updates.collect { nameInBranch ->
|
dataSet.updates().collect {
|
||||||
data(name + nameInBranch, dataSet.get(nameInBranch))
|
emit(name + it.name, it.data)
|
||||||
}
|
}
|
||||||
}
|
}
|
@ -29,7 +29,7 @@ internal class ActionsTest {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
fun testDynamicMapAction() = runTest {
|
fun testDynamicMapAction() = runTest {
|
||||||
val data: DataSourceBuilder<Int> = DataSource(this)
|
val data: MutableDataTree<Int> = MutableDataTree(this)
|
||||||
|
|
||||||
val plusOne = Action.mapping<Int, Int> {
|
val plusOne = Action.mapping<Int, Int> {
|
||||||
result { it + 1 }
|
result { it + 1 }
|
||||||
@ -44,7 +44,6 @@ internal class ActionsTest {
|
|||||||
delay(20)
|
delay(20)
|
||||||
|
|
||||||
assertEquals(2, result["1"]?.await())
|
assertEquals(2, result["1"]?.await())
|
||||||
data.close()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
@ -7,7 +7,7 @@ import kotlin.test.Test
|
|||||||
import kotlin.test.assertEquals
|
import kotlin.test.assertEquals
|
||||||
|
|
||||||
|
|
||||||
internal class DataTreeBuilderTest {
|
internal class LegacyGenericDataTreeBuilderTest {
|
||||||
@Test
|
@Test
|
||||||
fun testTreeBuild() = runBlocking {
|
fun testTreeBuild() = runBlocking {
|
||||||
val node = DataTree<Any> {
|
val node = DataTree<Any> {
|
||||||
@ -29,7 +29,7 @@ internal class DataTreeBuilderTest {
|
|||||||
@OptIn(DFExperimental::class)
|
@OptIn(DFExperimental::class)
|
||||||
@Test
|
@Test
|
||||||
fun testDataUpdate() = runBlocking {
|
fun testDataUpdate() = runBlocking {
|
||||||
val updateData: DataTree<Any> = DataTree {
|
val updateData = DataTree<Any> {
|
||||||
"update" put {
|
"update" put {
|
||||||
"a" put Data.static("a")
|
"a" put Data.static("a")
|
||||||
"b" put Data.static("b")
|
"b" put Data.static("b")
|
||||||
@ -56,7 +56,7 @@ internal class DataTreeBuilderTest {
|
|||||||
try {
|
try {
|
||||||
lateinit var updateJob: Job
|
lateinit var updateJob: Job
|
||||||
supervisorScope {
|
supervisorScope {
|
||||||
val subNode = DataSource<Int>(this) {
|
val subNode = ObservableDataTree<Int>(this) {
|
||||||
updateJob = launch {
|
updateJob = launch {
|
||||||
repeat(10) {
|
repeat(10) {
|
||||||
delay(10)
|
delay(10)
|
||||||
@ -66,16 +66,16 @@ internal class DataTreeBuilderTest {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
launch {
|
launch {
|
||||||
subNode.updatesWithData.collect {
|
subNode.updates().collect {
|
||||||
println(it)
|
println(it)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
val rootNode = DataSource<Int>(this) {
|
val rootNode = ObservableDataTree<Int>(this) {
|
||||||
setAndWatch("sub".asName(), subNode)
|
setAndWatch("sub".asName(), subNode)
|
||||||
}
|
}
|
||||||
|
|
||||||
launch {
|
launch {
|
||||||
rootNode.updatesWithData.collect {
|
rootNode.updates().collect {
|
||||||
println(it)
|
println(it)
|
||||||
}
|
}
|
||||||
}
|
}
|
@ -17,11 +17,7 @@ import kotlin.reflect.typeOf
|
|||||||
/**
|
/**
|
||||||
* Reader of a custom object from input
|
* Reader of a custom object from input
|
||||||
*/
|
*/
|
||||||
public interface IOReader<out T> {
|
public fun interface IOReader<out T> {
|
||||||
/**
|
|
||||||
* The type of object being read
|
|
||||||
*/
|
|
||||||
public val type: KType
|
|
||||||
|
|
||||||
public fun readFrom(source: Source): T
|
public fun readFrom(source: Source): T
|
||||||
|
|
||||||
@ -32,7 +28,6 @@ public interface IOReader<out T> {
|
|||||||
* no-op reader for binaries.
|
* no-op reader for binaries.
|
||||||
*/
|
*/
|
||||||
public val binary: IOReader<Binary> = object : IOReader<Binary> {
|
public val binary: IOReader<Binary> = object : IOReader<Binary> {
|
||||||
override val type: KType = typeOf<Binary>()
|
|
||||||
|
|
||||||
override fun readFrom(source: Source): Binary = source.readByteArray().asBinary()
|
override fun readFrom(source: Source): Binary = source.readByteArray().asBinary()
|
||||||
|
|
||||||
@ -42,8 +37,6 @@ public interface IOReader<out T> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public inline fun <reified T> IOReader(crossinline read: Source.() -> T): IOReader<T> = object : IOReader<T> {
|
public inline fun <reified T> IOReader(crossinline read: Source.() -> T): IOReader<T> = object : IOReader<T> {
|
||||||
override val type: KType = typeOf<T>()
|
|
||||||
|
|
||||||
override fun readFrom(source: Source): T = source.read()
|
override fun readFrom(source: Source): T = source.read()
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -61,19 +54,19 @@ public fun <T : Any> Source.readWith(format: IOReader<T>): T = format.readFrom(t
|
|||||||
/**
|
/**
|
||||||
* Read given binary as an object using given format
|
* Read given binary as an object using given format
|
||||||
*/
|
*/
|
||||||
public fun <T : Any> Binary.readWith(format: IOReader<T>): T = read {
|
public fun <T> Binary.readWith(format: IOReader<T>): T = read {
|
||||||
readWith(format)
|
readWith(format)
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Write an object to the [Sink] with given [format]
|
* Write an object to the [Sink] with given [format]
|
||||||
*/
|
*/
|
||||||
public fun <T : Any> Sink.writeWith(format: IOWriter<T>, obj: T): Unit =
|
public fun <T> Sink.writeWith(format: IOWriter<T>, obj: T): Unit =
|
||||||
format.writeTo(this, obj)
|
format.writeTo(this, obj)
|
||||||
|
|
||||||
|
|
||||||
@DfType(IO_FORMAT_TYPE)
|
@DfType(IO_FORMAT_TYPE)
|
||||||
public interface IOFormatFactory<T : Any> : Factory<IOFormat<T>>, Named {
|
public interface IOFormatFactory<T> : Factory<IOFormat<T>>, Named {
|
||||||
/**
|
/**
|
||||||
* Explicit type for dynamic type checks
|
* Explicit type for dynamic type checks
|
||||||
*/
|
*/
|
||||||
@ -86,7 +79,7 @@ public interface IOFormatFactory<T : Any> : Factory<IOFormat<T>>, Named {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public fun <T : Any> Binary(obj: T, format: IOWriter<T>): Binary = Binary { format.writeTo(this, obj) }
|
public fun <T> Binary(obj: T, format: IOWriter<T>): Binary = Binary { format.writeTo(this, obj) }
|
||||||
|
|
||||||
public object FloatIOFormat : IOFormat<Float>, IOFormatFactory<Float> {
|
public object FloatIOFormat : IOFormat<Float>, IOFormatFactory<Float> {
|
||||||
override fun build(context: Context, meta: Meta): IOFormat<Float> = this
|
override fun build(context: Context, meta: Meta): IOFormat<Float> = this
|
||||||
|
@ -5,7 +5,6 @@ import space.kscience.dataforge.io.EnvelopeFormatFactory.Companion.ENVELOPE_FORM
|
|||||||
import space.kscience.dataforge.io.IOFormatFactory.Companion.IO_FORMAT_TYPE
|
import space.kscience.dataforge.io.IOFormatFactory.Companion.IO_FORMAT_TYPE
|
||||||
import space.kscience.dataforge.io.MetaFormatFactory.Companion.META_FORMAT_TYPE
|
import space.kscience.dataforge.io.MetaFormatFactory.Companion.META_FORMAT_TYPE
|
||||||
import space.kscience.dataforge.meta.Meta
|
import space.kscience.dataforge.meta.Meta
|
||||||
import space.kscience.dataforge.meta.get
|
|
||||||
import space.kscience.dataforge.meta.string
|
import space.kscience.dataforge.meta.string
|
||||||
import space.kscience.dataforge.misc.DFInternal
|
import space.kscience.dataforge.misc.DFInternal
|
||||||
import space.kscience.dataforge.names.Name
|
import space.kscience.dataforge.names.Name
|
||||||
@ -21,11 +20,11 @@ public class IOPlugin(meta: Meta) : AbstractPlugin(meta) {
|
|||||||
|
|
||||||
@Suppress("UNCHECKED_CAST")
|
@Suppress("UNCHECKED_CAST")
|
||||||
@DFInternal
|
@DFInternal
|
||||||
public fun <T : Any> resolveIOFormat(type: KType, meta: Meta): IOFormat<T>? =
|
public fun <T> resolveIOFormat(type: KType, meta: Meta): IOFormat<T>? =
|
||||||
ioFormatFactories.singleOrNull { it.type == type }?.build(context, meta) as? IOFormat<T>
|
ioFormatFactories.singleOrNull { it.type == type }?.build(context, meta) as? IOFormat<T>
|
||||||
|
|
||||||
@OptIn(DFInternal::class)
|
@OptIn(DFInternal::class)
|
||||||
public inline fun <reified T : Any> resolveIOFormat(meta: Meta = Meta.EMPTY): IOFormat<T>? =
|
public inline fun <reified T> resolveIOFormat(meta: Meta = Meta.EMPTY): IOFormat<T>? =
|
||||||
resolveIOFormat(typeOf<T>(), meta)
|
resolveIOFormat(typeOf<T>(), meta)
|
||||||
|
|
||||||
|
|
||||||
|
@ -6,8 +6,6 @@ import kotlinx.serialization.json.encodeToJsonElement
|
|||||||
import kotlinx.serialization.serializer
|
import kotlinx.serialization.serializer
|
||||||
import space.kscience.dataforge.meta.descriptors.MetaDescriptor
|
import space.kscience.dataforge.meta.descriptors.MetaDescriptor
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
import space.kscience.dataforge.misc.DFExperimental
|
||||||
import kotlin.reflect.KType
|
|
||||||
import kotlin.reflect.typeOf
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -15,15 +13,10 @@ import kotlin.reflect.typeOf
|
|||||||
*/
|
*/
|
||||||
public interface MetaConverter<T>: MetaSpec<T> {
|
public interface MetaConverter<T>: MetaSpec<T> {
|
||||||
|
|
||||||
/**
|
|
||||||
* Runtime type of [T]
|
|
||||||
*/
|
|
||||||
public val type: KType
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A descriptor for resulting meta
|
* A descriptor for resulting meta
|
||||||
*/
|
*/
|
||||||
override val descriptor: MetaDescriptor get() = MetaDescriptor.EMPTY
|
override val descriptor: MetaDescriptor? get() = null
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Attempt conversion of [source] to an object or return null if conversion failed
|
* Attempt conversion of [source] to an object or return null if conversion failed
|
||||||
@ -38,22 +31,16 @@ public interface MetaConverter<T>: MetaSpec<T> {
|
|||||||
public companion object {
|
public companion object {
|
||||||
|
|
||||||
public val meta: MetaConverter<Meta> = object : MetaConverter<Meta> {
|
public val meta: MetaConverter<Meta> = object : MetaConverter<Meta> {
|
||||||
override val type: KType = typeOf<Meta>()
|
|
||||||
|
|
||||||
override fun readOrNull(source: Meta): Meta = source
|
override fun readOrNull(source: Meta): Meta = source
|
||||||
override fun convert(obj: Meta): Meta = obj
|
override fun convert(obj: Meta): Meta = obj
|
||||||
}
|
}
|
||||||
|
|
||||||
public val value: MetaConverter<Value> = object : MetaConverter<Value> {
|
public val value: MetaConverter<Value> = object : MetaConverter<Value> {
|
||||||
override val type: KType = typeOf<Value>()
|
|
||||||
|
|
||||||
override fun readOrNull(source: Meta): Value? = source.value
|
override fun readOrNull(source: Meta): Value? = source.value
|
||||||
override fun convert(obj: Value): Meta = Meta(obj)
|
override fun convert(obj: Value): Meta = Meta(obj)
|
||||||
}
|
}
|
||||||
|
|
||||||
public val string: MetaConverter<String> = object : MetaConverter<String> {
|
public val string: MetaConverter<String> = object : MetaConverter<String> {
|
||||||
override val type: KType = typeOf<String>()
|
|
||||||
|
|
||||||
override val descriptor: MetaDescriptor = MetaDescriptor {
|
override val descriptor: MetaDescriptor = MetaDescriptor {
|
||||||
valueType(ValueType.STRING)
|
valueType(ValueType.STRING)
|
||||||
}
|
}
|
||||||
@ -64,8 +51,6 @@ public interface MetaConverter<T>: MetaSpec<T> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public val boolean: MetaConverter<Boolean> = object : MetaConverter<Boolean> {
|
public val boolean: MetaConverter<Boolean> = object : MetaConverter<Boolean> {
|
||||||
override val type: KType = typeOf<Boolean>()
|
|
||||||
|
|
||||||
override val descriptor: MetaDescriptor = MetaDescriptor {
|
override val descriptor: MetaDescriptor = MetaDescriptor {
|
||||||
valueType(ValueType.BOOLEAN)
|
valueType(ValueType.BOOLEAN)
|
||||||
}
|
}
|
||||||
@ -75,8 +60,6 @@ public interface MetaConverter<T>: MetaSpec<T> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public val number: MetaConverter<Number> = object : MetaConverter<Number> {
|
public val number: MetaConverter<Number> = object : MetaConverter<Number> {
|
||||||
override val type: KType = typeOf<Number>()
|
|
||||||
|
|
||||||
override val descriptor: MetaDescriptor = MetaDescriptor {
|
override val descriptor: MetaDescriptor = MetaDescriptor {
|
||||||
valueType(ValueType.NUMBER)
|
valueType(ValueType.NUMBER)
|
||||||
}
|
}
|
||||||
@ -86,8 +69,6 @@ public interface MetaConverter<T>: MetaSpec<T> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public val double: MetaConverter<Double> = object : MetaConverter<Double> {
|
public val double: MetaConverter<Double> = object : MetaConverter<Double> {
|
||||||
override val type: KType = typeOf<Double>()
|
|
||||||
|
|
||||||
override val descriptor: MetaDescriptor = MetaDescriptor {
|
override val descriptor: MetaDescriptor = MetaDescriptor {
|
||||||
valueType(ValueType.NUMBER)
|
valueType(ValueType.NUMBER)
|
||||||
}
|
}
|
||||||
@ -97,8 +78,6 @@ public interface MetaConverter<T>: MetaSpec<T> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public val float: MetaConverter<Float> = object : MetaConverter<Float> {
|
public val float: MetaConverter<Float> = object : MetaConverter<Float> {
|
||||||
override val type: KType = typeOf<Float>()
|
|
||||||
|
|
||||||
override val descriptor: MetaDescriptor = MetaDescriptor {
|
override val descriptor: MetaDescriptor = MetaDescriptor {
|
||||||
valueType(ValueType.NUMBER)
|
valueType(ValueType.NUMBER)
|
||||||
}
|
}
|
||||||
@ -108,8 +87,6 @@ public interface MetaConverter<T>: MetaSpec<T> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public val int: MetaConverter<Int> = object : MetaConverter<Int> {
|
public val int: MetaConverter<Int> = object : MetaConverter<Int> {
|
||||||
override val type: KType = typeOf<Int>()
|
|
||||||
|
|
||||||
override val descriptor: MetaDescriptor = MetaDescriptor {
|
override val descriptor: MetaDescriptor = MetaDescriptor {
|
||||||
valueType(ValueType.NUMBER)
|
valueType(ValueType.NUMBER)
|
||||||
}
|
}
|
||||||
@ -119,8 +96,6 @@ public interface MetaConverter<T>: MetaSpec<T> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public val long: MetaConverter<Long> = object : MetaConverter<Long> {
|
public val long: MetaConverter<Long> = object : MetaConverter<Long> {
|
||||||
override val type: KType = typeOf<Long>()
|
|
||||||
|
|
||||||
override val descriptor: MetaDescriptor = MetaDescriptor {
|
override val descriptor: MetaDescriptor = MetaDescriptor {
|
||||||
valueType(ValueType.NUMBER)
|
valueType(ValueType.NUMBER)
|
||||||
}
|
}
|
||||||
@ -130,8 +105,6 @@ public interface MetaConverter<T>: MetaSpec<T> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public inline fun <reified E : Enum<E>> enum(): MetaConverter<E> = object : MetaConverter<E> {
|
public inline fun <reified E : Enum<E>> enum(): MetaConverter<E> = object : MetaConverter<E> {
|
||||||
override val type: KType = typeOf<E>()
|
|
||||||
|
|
||||||
override val descriptor: MetaDescriptor = MetaDescriptor {
|
override val descriptor: MetaDescriptor = MetaDescriptor {
|
||||||
valueType(ValueType.STRING)
|
valueType(ValueType.STRING)
|
||||||
allowedValues(enumValues<E>())
|
allowedValues(enumValues<E>())
|
||||||
@ -147,8 +120,6 @@ public interface MetaConverter<T>: MetaSpec<T> {
|
|||||||
writer: (T) -> Value = { Value.of(it) },
|
writer: (T) -> Value = { Value.of(it) },
|
||||||
reader: (Value) -> T,
|
reader: (Value) -> T,
|
||||||
): MetaConverter<List<T>> = object : MetaConverter<List<T>> {
|
): MetaConverter<List<T>> = object : MetaConverter<List<T>> {
|
||||||
override val type: KType = typeOf<List<T>>()
|
|
||||||
|
|
||||||
override val descriptor: MetaDescriptor = MetaDescriptor {
|
override val descriptor: MetaDescriptor = MetaDescriptor {
|
||||||
valueType(ValueType.LIST)
|
valueType(ValueType.LIST)
|
||||||
}
|
}
|
||||||
@ -165,7 +136,6 @@ public interface MetaConverter<T>: MetaSpec<T> {
|
|||||||
public inline fun <reified T> serializable(
|
public inline fun <reified T> serializable(
|
||||||
descriptor: MetaDescriptor? = null,
|
descriptor: MetaDescriptor? = null,
|
||||||
): MetaConverter<T> = object : MetaConverter<T> {
|
): MetaConverter<T> = object : MetaConverter<T> {
|
||||||
override val type: KType = typeOf<T>()
|
|
||||||
private val serializer: KSerializer<T> = serializer()
|
private val serializer: KSerializer<T> = serializer()
|
||||||
|
|
||||||
override fun readOrNull(source: Meta): T? {
|
override fun readOrNull(source: Meta): T? {
|
||||||
|
@ -166,9 +166,9 @@ public inline fun <T : Scheme> T.copy(spec: SchemeSpec<T>, block: T.() -> Unit =
|
|||||||
/**
|
/**
|
||||||
* A specification for simplified generation of wrappers
|
* A specification for simplified generation of wrappers
|
||||||
*/
|
*/
|
||||||
public open class SchemeSpec<out T : Scheme>(
|
public open class SchemeSpec<T : Scheme>(
|
||||||
private val builder: () -> T,
|
private val builder: () -> T,
|
||||||
) : MetaSpec<T> {
|
) : MetaConverter<T> {
|
||||||
|
|
||||||
override val descriptor: MetaDescriptor? get() = null
|
override val descriptor: MetaDescriptor? get() = null
|
||||||
|
|
||||||
@ -187,6 +187,8 @@ public open class SchemeSpec<out T : Scheme>(
|
|||||||
it.initialize(MutableMeta(), Meta.EMPTY, descriptor)
|
it.initialize(MutableMeta(), Meta.EMPTY, descriptor)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
override fun convert(obj: T): Meta = obj.meta
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A convenience method to use specifications in builders
|
* A convenience method to use specifications in builders
|
||||||
*/
|
*/
|
||||||
|
@ -113,6 +113,13 @@ public class Name(public val tokens: List<NameToken>) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Transform this [Name] to a string without escaping special characters in tokens.
|
||||||
|
*
|
||||||
|
* Parsing it back will produce a valid, but different name
|
||||||
|
*/
|
||||||
|
public fun Name.toStringUnescaped(): String = tokens.joinToString(separator = Name.NAME_SEPARATOR) { it.toStringUnescaped() }
|
||||||
|
|
||||||
public operator fun Name.get(i: Int): NameToken = tokens[i]
|
public operator fun Name.get(i: Int): NameToken = tokens[i]
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -1,46 +0,0 @@
|
|||||||
package space.kscience.dataforge.workspace
|
|
||||||
|
|
||||||
import space.kscience.dataforge.data.DataTree.Companion.META_ITEM_NAME_TOKEN
|
|
||||||
import space.kscience.dataforge.io.Envelope
|
|
||||||
import space.kscience.dataforge.io.IOReader
|
|
||||||
import space.kscience.dataforge.meta.Meta
|
|
||||||
import space.kscience.dataforge.meta.descriptors.MetaDescriptor
|
|
||||||
import space.kscience.dataforge.names.Name
|
|
||||||
import space.kscience.dataforge.names.asName
|
|
||||||
import kotlin.reflect.KType
|
|
||||||
|
|
||||||
public abstract class EnvelopeTask<T : Any>(
|
|
||||||
override val descriptor: MetaDescriptor?,
|
|
||||||
private val reader: IOReader<T>,
|
|
||||||
) : Task<T> {
|
|
||||||
|
|
||||||
public abstract suspend fun produceEnvelopes(
|
|
||||||
workspace: Workspace,
|
|
||||||
taskName: Name,
|
|
||||||
taskMeta: Meta,
|
|
||||||
): Map<Name, Envelope>
|
|
||||||
|
|
||||||
override suspend fun execute(workspace: Workspace, taskName: Name, taskMeta: Meta): TaskResult<T> =
|
|
||||||
Result(workspace, taskName, taskMeta, reader, produceEnvelopes(workspace, taskName, taskMeta))
|
|
||||||
|
|
||||||
private class Result<T : Any>(
|
|
||||||
override val workspace: Workspace,
|
|
||||||
override val taskName: Name,
|
|
||||||
override val taskMeta: Meta,
|
|
||||||
val reader: IOReader<T>,
|
|
||||||
envelopes: Map<Name, Envelope>,
|
|
||||||
) : TaskResult<T> {
|
|
||||||
|
|
||||||
private val dataMap = envelopes.mapValues {
|
|
||||||
workspace.wrapData(it.value.toData(reader), it.key, taskName, taskMeta)
|
|
||||||
}
|
|
||||||
override val meta: Meta get() = dataMap[META_ITEM_NAME_TOKEN.asName()]?.meta ?: Meta.EMPTY
|
|
||||||
|
|
||||||
override val dataType: KType get() = reader.type
|
|
||||||
|
|
||||||
override fun iterator(): Iterator<TaskData<T>> = dataMap.values.iterator()
|
|
||||||
|
|
||||||
override fun get(name: Name): TaskData<T>? = dataMap[name]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
@ -1,9 +1,9 @@
|
|||||||
package space.kscience.dataforge.workspace
|
package space.kscience.dataforge.workspace
|
||||||
|
|
||||||
import kotlinx.coroutines.withContext
|
import kotlinx.coroutines.withContext
|
||||||
import space.kscience.dataforge.data.DataSetBuilder
|
import space.kscience.dataforge.data.DataSink
|
||||||
import space.kscience.dataforge.data.DataTree
|
|
||||||
import space.kscience.dataforge.data.GoalExecutionRestriction
|
import space.kscience.dataforge.data.GoalExecutionRestriction
|
||||||
|
import space.kscience.dataforge.data.MutableDataTree
|
||||||
import space.kscience.dataforge.meta.Meta
|
import space.kscience.dataforge.meta.Meta
|
||||||
import space.kscience.dataforge.meta.MetaRepr
|
import space.kscience.dataforge.meta.MetaRepr
|
||||||
import space.kscience.dataforge.meta.MetaSpec
|
import space.kscience.dataforge.meta.MetaSpec
|
||||||
@ -20,7 +20,7 @@ import kotlin.reflect.typeOf
|
|||||||
* In general no computations should be made until the result is called.
|
* In general no computations should be made until the result is called.
|
||||||
*/
|
*/
|
||||||
@DfType(TYPE)
|
@DfType(TYPE)
|
||||||
public interface Task<out T : Any> : Described {
|
public interface Task<T> : Described {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A task identification string used to compare tasks and check task body for change
|
* A task identification string used to compare tasks and check task body for change
|
||||||
@ -45,7 +45,7 @@ public interface Task<out T : Any> : Described {
|
|||||||
/**
|
/**
|
||||||
* A [Task] with [MetaSpec] for wrapping and unwrapping task configuration
|
* A [Task] with [MetaSpec] for wrapping and unwrapping task configuration
|
||||||
*/
|
*/
|
||||||
public interface TaskWithSpec<out T : Any, C : Any> : Task<T> {
|
public interface TaskWithSpec<T, C : Any> : Task<T> {
|
||||||
public val spec: MetaSpec<C>
|
public val spec: MetaSpec<C>
|
||||||
override val descriptor: MetaDescriptor? get() = spec.descriptor
|
override val descriptor: MetaDescriptor? get() = spec.descriptor
|
||||||
|
|
||||||
@ -61,12 +61,12 @@ public interface TaskWithSpec<out T : Any, C : Any> : Task<T> {
|
|||||||
// block: C.() -> Unit = {},
|
// block: C.() -> Unit = {},
|
||||||
//): TaskResult<T> = execute(workspace, taskName, spec(block))
|
//): TaskResult<T> = execute(workspace, taskName, spec(block))
|
||||||
|
|
||||||
public class TaskResultBuilder<in T : Any>(
|
public class TaskResultBuilder<T>(
|
||||||
public val workspace: Workspace,
|
public val workspace: Workspace,
|
||||||
public val taskName: Name,
|
public val taskName: Name,
|
||||||
public val taskMeta: Meta,
|
public val taskMeta: Meta,
|
||||||
private val dataDrop: DataSetBuilder<T>,
|
private val dataSink: DataSink<T>,
|
||||||
) : DataSetBuilder<T> by dataDrop
|
) : DataSink<T> by dataSink
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a [Task] that composes a result using [builder]. Only data from the workspace could be used.
|
* Create a [Task] that composes a result using [builder]. Only data from the workspace could be used.
|
||||||
@ -90,7 +90,7 @@ public fun <T : Any> Task(
|
|||||||
taskMeta: Meta,
|
taskMeta: Meta,
|
||||||
): TaskResult<T> = withContext(GoalExecutionRestriction() + workspace.goalLogger) {
|
): TaskResult<T> = withContext(GoalExecutionRestriction() + workspace.goalLogger) {
|
||||||
//TODO use safe builder and check for external data on add and detects cycles
|
//TODO use safe builder and check for external data on add and detects cycles
|
||||||
val dataset = DataTree<T>(resultType) {
|
val dataset = MutableDataTree<T>(resultType, this).apply {
|
||||||
TaskResultBuilder(workspace, taskName, taskMeta, this).apply { builder() }
|
TaskResultBuilder(workspace, taskName, taskMeta, this).apply { builder() }
|
||||||
}
|
}
|
||||||
workspace.wrapResult(dataset, taskName, taskMeta)
|
workspace.wrapResult(dataset, taskName, taskMeta)
|
||||||
@ -111,6 +111,7 @@ public inline fun <reified T : Any> Task(
|
|||||||
* @param specification a specification for task configuration
|
* @param specification a specification for task configuration
|
||||||
* @param builder for resulting data set
|
* @param builder for resulting data set
|
||||||
*/
|
*/
|
||||||
|
|
||||||
@Suppress("FunctionName")
|
@Suppress("FunctionName")
|
||||||
public fun <T : Any, C : MetaRepr> Task(
|
public fun <T : Any, C : MetaRepr> Task(
|
||||||
resultType: KType,
|
resultType: KType,
|
||||||
@ -126,7 +127,7 @@ public fun <T : Any, C : MetaRepr> Task(
|
|||||||
): TaskResult<T> = withContext(GoalExecutionRestriction() + workspace.goalLogger) {
|
): TaskResult<T> = withContext(GoalExecutionRestriction() + workspace.goalLogger) {
|
||||||
//TODO use safe builder and check for external data on add and detects cycles
|
//TODO use safe builder and check for external data on add and detects cycles
|
||||||
val taskMeta = configuration.toMeta()
|
val taskMeta = configuration.toMeta()
|
||||||
val dataset = DataTree<T>(resultType) {
|
val dataset = MutableDataTree<T>(resultType, this).apply {
|
||||||
TaskResultBuilder(workspace, taskName, taskMeta, this).apply { builder(configuration) }
|
TaskResultBuilder(workspace, taskName, taskMeta, this).apply { builder(configuration) }
|
||||||
}
|
}
|
||||||
workspace.wrapResult(dataset, taskName, taskMeta)
|
workspace.wrapResult(dataset, taskName, taskMeta)
|
||||||
|
@ -1,50 +0,0 @@
|
|||||||
package space.kscience.dataforge.workspace
|
|
||||||
|
|
||||||
import space.kscience.dataforge.data.Data
|
|
||||||
import space.kscience.dataforge.data.NamedData
|
|
||||||
import space.kscience.dataforge.meta.Meta
|
|
||||||
import space.kscience.dataforge.names.Name
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A [Workspace]-locked [NamedData], that serves as a computation model.
|
|
||||||
*/
|
|
||||||
public interface TaskData<out T : Any> : NamedData<T> {
|
|
||||||
/**
|
|
||||||
* The [Workspace] this data belongs to
|
|
||||||
*/
|
|
||||||
public val workspace: Workspace
|
|
||||||
|
|
||||||
/**
|
|
||||||
* The name of the stage that produced this data. [Name.EMPTY] if the workspace intrinsic data is used.
|
|
||||||
*/
|
|
||||||
public val taskName: Name
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Stage configuration used to produce this data.
|
|
||||||
*/
|
|
||||||
public val taskMeta: Meta
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Dependencies that allow to compute transitive dependencies as well.
|
|
||||||
*/
|
|
||||||
// override val dependencies: Collection<TaskData<*>>
|
|
||||||
}
|
|
||||||
|
|
||||||
private class TaskDataImpl<out T : Any>(
|
|
||||||
override val workspace: Workspace,
|
|
||||||
override val data: Data<T>,
|
|
||||||
override val name: Name,
|
|
||||||
override val taskName: Name,
|
|
||||||
override val taskMeta: Meta,
|
|
||||||
) : TaskData<T>, Data<T> by data {
|
|
||||||
// override val dependencies: Collection<TaskData<*>> = data.dependencies.map {
|
|
||||||
// it as? TaskData<*> ?: error("TaskData can't depend on external data")
|
|
||||||
// }
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Adopt data into this workspace
|
|
||||||
*/
|
|
||||||
public fun <T : Any> Workspace.wrapData(data: Data<T>, name: Name, taskName: Name, taskMeta: Meta): TaskData<T> =
|
|
||||||
TaskDataImpl(this, data, name, taskName, taskMeta)
|
|
||||||
|
|
@ -1,54 +1,27 @@
|
|||||||
package space.kscience.dataforge.workspace
|
package space.kscience.dataforge.workspace
|
||||||
|
|
||||||
import space.kscience.dataforge.data.DataSet
|
import space.kscience.dataforge.data.ObservableDataTree
|
||||||
import space.kscience.dataforge.data.forEach
|
|
||||||
import space.kscience.dataforge.meta.Meta
|
import space.kscience.dataforge.meta.Meta
|
||||||
import space.kscience.dataforge.names.Name
|
import space.kscience.dataforge.names.Name
|
||||||
|
import kotlin.reflect.KType
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A result of a [Task]
|
* A result of a [Task]
|
||||||
|
* @param workspace the [Workspace] that produced the result
|
||||||
|
* @param taskName the name of the task that produced the result
|
||||||
|
* @param taskMeta The configuration of the task that produced the result
|
||||||
*/
|
*/
|
||||||
public interface TaskResult<out T : Any> : DataSet<T> {
|
public data class TaskResult<T>(
|
||||||
/**
|
public val data: ObservableDataTree<T>,
|
||||||
* The [Workspace] this [DataSet] belongs to
|
public val workspace: Workspace,
|
||||||
*/
|
public val taskName: Name,
|
||||||
public val workspace: Workspace
|
public val taskMeta: Meta,
|
||||||
|
) {
|
||||||
/**
|
val dataType: KType get() = data.dataType
|
||||||
* The [Name] of the stage that produced this [DataSet]
|
|
||||||
*/
|
|
||||||
public val taskName: Name
|
|
||||||
|
|
||||||
/**
|
|
||||||
* The configuration of the stage that produced this [DataSet]
|
|
||||||
*/
|
|
||||||
public val taskMeta: Meta
|
|
||||||
|
|
||||||
override fun iterator(): Iterator<TaskData<T>>
|
|
||||||
|
|
||||||
override fun get(name: Name): TaskData<T>?
|
|
||||||
}
|
|
||||||
|
|
||||||
private class TaskResultImpl<out T : Any>(
|
|
||||||
override val workspace: Workspace,
|
|
||||||
override val taskName: Name,
|
|
||||||
override val taskMeta: Meta,
|
|
||||||
val dataSet: DataSet<T>,
|
|
||||||
) : TaskResult<T>, DataSet<T> by dataSet {
|
|
||||||
|
|
||||||
override fun iterator(): Iterator<TaskData<T>> = iterator {
|
|
||||||
dataSet.forEach {
|
|
||||||
yield(workspace.wrapData(it, it.name, taskName, taskMeta))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
override fun get(name: Name): TaskData<T>? = dataSet[name]?.let {
|
|
||||||
workspace.wrapData(it, name, taskName, taskMeta)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Wrap data into [TaskResult]
|
* Wrap data into [TaskResult]
|
||||||
*/
|
*/
|
||||||
public fun <T : Any> Workspace.wrapResult(dataSet: DataSet<T>, taskName: Name, taskMeta: Meta): TaskResult<T> =
|
public fun <T> Workspace.wrapResult(data: ObservableDataTree<T>, taskName: Name, taskMeta: Meta): TaskResult<T> =
|
||||||
TaskResultImpl(this, taskName, taskMeta, dataSet)
|
TaskResult(data, this, taskName, taskMeta)
|
@ -1,29 +1,32 @@
|
|||||||
package space.kscience.dataforge.workspace
|
package space.kscience.dataforge.workspace
|
||||||
|
|
||||||
|
import kotlinx.coroutines.CoroutineScope
|
||||||
import space.kscience.dataforge.context.ContextAware
|
import space.kscience.dataforge.context.ContextAware
|
||||||
import space.kscience.dataforge.data.Data
|
import space.kscience.dataforge.data.*
|
||||||
import space.kscience.dataforge.data.DataSet
|
|
||||||
import space.kscience.dataforge.data.asSequence
|
|
||||||
import space.kscience.dataforge.meta.Meta
|
import space.kscience.dataforge.meta.Meta
|
||||||
import space.kscience.dataforge.meta.MutableMeta
|
import space.kscience.dataforge.meta.MutableMeta
|
||||||
import space.kscience.dataforge.misc.DfType
|
import space.kscience.dataforge.misc.DfType
|
||||||
import space.kscience.dataforge.names.Name
|
import space.kscience.dataforge.names.Name
|
||||||
import space.kscience.dataforge.provider.Provider
|
import space.kscience.dataforge.provider.Provider
|
||||||
|
import kotlin.coroutines.CoroutineContext
|
||||||
|
|
||||||
|
|
||||||
public interface DataSelector<T: Any>{
|
public fun interface DataSelector<T> {
|
||||||
public suspend fun select(workspace: Workspace, meta: Meta): DataSet<T>
|
public suspend fun select(workspace: Workspace, meta: Meta): DataTree<T>
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* An environment for pull-mode computation
|
* An environment for pull-mode computation
|
||||||
*/
|
*/
|
||||||
@DfType(Workspace.TYPE)
|
@DfType(Workspace.TYPE)
|
||||||
public interface Workspace : ContextAware, Provider {
|
public interface Workspace : ContextAware, Provider, CoroutineScope {
|
||||||
|
|
||||||
|
override val coroutineContext: CoroutineContext get() = context.coroutineContext
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The whole data node for current workspace
|
* The whole data node for current workspace
|
||||||
*/
|
*/
|
||||||
public val data: TaskResult<*>
|
public val data: ObservableDataTree<*>
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* All targets associated with the workspace
|
* All targets associated with the workspace
|
||||||
@ -49,8 +52,8 @@ public interface Workspace : ContextAware, Provider {
|
|||||||
return task.execute(this, taskName, taskMeta)
|
return task.execute(this, taskName, taskMeta)
|
||||||
}
|
}
|
||||||
|
|
||||||
public suspend fun produceData(taskName: Name, taskMeta: Meta, name: Name): TaskData<*>? =
|
public suspend fun produceData(taskName: Name, taskMeta: Meta, name: Name): Data<*>? =
|
||||||
produce(taskName, taskMeta)[name]
|
produce(taskName, taskMeta).data[name]
|
||||||
|
|
||||||
public companion object {
|
public companion object {
|
||||||
public const val TYPE: String = "workspace"
|
public const val TYPE: String = "workspace"
|
||||||
|
@ -5,9 +5,9 @@ import space.kscience.dataforge.actions.Action
|
|||||||
import space.kscience.dataforge.context.Context
|
import space.kscience.dataforge.context.Context
|
||||||
import space.kscience.dataforge.context.ContextBuilder
|
import space.kscience.dataforge.context.ContextBuilder
|
||||||
import space.kscience.dataforge.context.Global
|
import space.kscience.dataforge.context.Global
|
||||||
import space.kscience.dataforge.data.DataSet
|
import space.kscience.dataforge.data.DataSink
|
||||||
import space.kscience.dataforge.data.DataSource
|
import space.kscience.dataforge.data.DataTree
|
||||||
import space.kscience.dataforge.data.DataSourceBuilder
|
import space.kscience.dataforge.data.MutableDataTree
|
||||||
import space.kscience.dataforge.meta.*
|
import space.kscience.dataforge.meta.*
|
||||||
import space.kscience.dataforge.meta.descriptors.MetaDescriptor
|
import space.kscience.dataforge.meta.descriptors.MetaDescriptor
|
||||||
import space.kscience.dataforge.meta.descriptors.MetaDescriptorBuilder
|
import space.kscience.dataforge.meta.descriptors.MetaDescriptorBuilder
|
||||||
@ -17,13 +17,14 @@ import space.kscience.dataforge.names.asName
|
|||||||
import kotlin.collections.set
|
import kotlin.collections.set
|
||||||
import kotlin.properties.PropertyDelegateProvider
|
import kotlin.properties.PropertyDelegateProvider
|
||||||
import kotlin.properties.ReadOnlyProperty
|
import kotlin.properties.ReadOnlyProperty
|
||||||
|
import kotlin.reflect.typeOf
|
||||||
|
|
||||||
public data class TaskReference<T : Any>(public val taskName: Name, public val task: Task<T>) : DataSelector<T> {
|
public data class TaskReference<T : Any>(public val taskName: Name, public val task: Task<T>) : DataSelector<T> {
|
||||||
|
|
||||||
@Suppress("UNCHECKED_CAST")
|
@Suppress("UNCHECKED_CAST")
|
||||||
override suspend fun select(workspace: Workspace, meta: Meta): DataSet<T> {
|
override suspend fun select(workspace: Workspace, meta: Meta): DataTree<T> {
|
||||||
if (workspace.tasks[taskName] == task) {
|
if (workspace.tasks[taskName] == task) {
|
||||||
return workspace.produce(taskName, meta) as TaskResult<T>
|
return workspace.produce(taskName, meta).data as DataTree<T>
|
||||||
} else {
|
} else {
|
||||||
error("Task $taskName does not belong to the workspace")
|
error("Task $taskName does not belong to the workspace")
|
||||||
}
|
}
|
||||||
@ -45,7 +46,7 @@ public inline fun <reified T : Any> TaskContainer.registerTask(
|
|||||||
): Unit = registerTask(Name.parse(name), Task(MetaDescriptor(descriptorBuilder), builder))
|
): Unit = registerTask(Name.parse(name), Task(MetaDescriptor(descriptorBuilder), builder))
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a new t
|
* Create and register a new task
|
||||||
*/
|
*/
|
||||||
public inline fun <reified T : Any> TaskContainer.buildTask(
|
public inline fun <reified T : Any> TaskContainer.buildTask(
|
||||||
name: String,
|
name: String,
|
||||||
@ -101,7 +102,7 @@ public inline fun <T : Any, reified R : Any> TaskContainer.action(
|
|||||||
noinline descriptorBuilder: MetaDescriptorBuilder.() -> Unit = {},
|
noinline descriptorBuilder: MetaDescriptorBuilder.() -> Unit = {},
|
||||||
): PropertyDelegateProvider<Any?, ReadOnlyProperty<Any?, TaskReference<R>>> =
|
): PropertyDelegateProvider<Any?, ReadOnlyProperty<Any?, TaskReference<R>>> =
|
||||||
task(MetaDescriptor(descriptorBuilder)) {
|
task(MetaDescriptor(descriptorBuilder)) {
|
||||||
result(action.execute(from(selector), taskMeta.copy(metaTransform)))
|
result(action.execute(workspace.context, from(selector), taskMeta.copy(metaTransform)))
|
||||||
}
|
}
|
||||||
|
|
||||||
public class WorkspaceBuilder(
|
public class WorkspaceBuilder(
|
||||||
@ -109,7 +110,7 @@ public class WorkspaceBuilder(
|
|||||||
private val coroutineScope: CoroutineScope = parentContext,
|
private val coroutineScope: CoroutineScope = parentContext,
|
||||||
) : TaskContainer {
|
) : TaskContainer {
|
||||||
private var context: Context? = null
|
private var context: Context? = null
|
||||||
private val data = DataSource<Any>(coroutineScope)
|
private val data = MutableDataTree<Any?>(typeOf<Any?>(), coroutineScope)
|
||||||
private val targets: HashMap<String, Meta> = HashMap()
|
private val targets: HashMap<String, Meta> = HashMap()
|
||||||
private val tasks = HashMap<Name, Task<*>>()
|
private val tasks = HashMap<Name, Task<*>>()
|
||||||
private var cache: WorkspaceCache? = null
|
private var cache: WorkspaceCache? = null
|
||||||
@ -124,7 +125,7 @@ public class WorkspaceBuilder(
|
|||||||
/**
|
/**
|
||||||
* Define intrinsic data for the workspace
|
* Define intrinsic data for the workspace
|
||||||
*/
|
*/
|
||||||
public fun data(builder: DataSourceBuilder<Any>.() -> Unit) {
|
public fun data(builder: DataSink<*>.() -> Unit) {
|
||||||
data.apply(builder)
|
data.apply(builder)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -149,7 +150,7 @@ public class WorkspaceBuilder(
|
|||||||
|
|
||||||
public fun build(): Workspace {
|
public fun build(): Workspace {
|
||||||
val postProcess: suspend (TaskResult<*>) -> TaskResult<*> = { result ->
|
val postProcess: suspend (TaskResult<*>) -> TaskResult<*> = { result ->
|
||||||
cache?.evaluate(result) ?: result
|
cache?.cache(result) ?: result
|
||||||
}
|
}
|
||||||
return WorkspaceImpl(context ?: parentContext, data, targets, tasks, postProcess)
|
return WorkspaceImpl(context ?: parentContext, data, targets, tasks, postProcess)
|
||||||
}
|
}
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
package space.kscience.dataforge.workspace
|
package space.kscience.dataforge.workspace
|
||||||
|
|
||||||
public interface WorkspaceCache {
|
public interface WorkspaceCache {
|
||||||
public suspend fun <T : Any> evaluate(result: TaskResult<T>): TaskResult<T>
|
public suspend fun <T> cache(result: TaskResult<T>): TaskResult<T>
|
||||||
}
|
}
|
||||||
|
@ -2,21 +2,19 @@ package space.kscience.dataforge.workspace
|
|||||||
|
|
||||||
import space.kscience.dataforge.context.Context
|
import space.kscience.dataforge.context.Context
|
||||||
import space.kscience.dataforge.context.gather
|
import space.kscience.dataforge.context.gather
|
||||||
import space.kscience.dataforge.data.DataSet
|
import space.kscience.dataforge.data.ObservableDataTree
|
||||||
import space.kscience.dataforge.meta.Meta
|
import space.kscience.dataforge.meta.Meta
|
||||||
import space.kscience.dataforge.names.Name
|
import space.kscience.dataforge.names.Name
|
||||||
|
|
||||||
|
|
||||||
internal class WorkspaceImpl internal constructor(
|
internal class WorkspaceImpl internal constructor(
|
||||||
override val context: Context,
|
override val context: Context,
|
||||||
data: DataSet<*>,
|
override val data: ObservableDataTree<*>,
|
||||||
override val targets: Map<String, Meta>,
|
override val targets: Map<String, Meta>,
|
||||||
tasks: Map<Name, Task<*>>,
|
tasks: Map<Name, Task<*>>,
|
||||||
private val postProcess: suspend (TaskResult<*>) -> TaskResult<*>,
|
private val postProcess: suspend (TaskResult<*>) -> TaskResult<*>,
|
||||||
) : Workspace {
|
) : Workspace {
|
||||||
|
|
||||||
override val data: TaskResult<*> = wrapResult(data, Name.EMPTY, Meta.EMPTY)
|
|
||||||
|
|
||||||
override val tasks: Map<Name, Task<*>> by lazy { context.gather<Task<*>>(Task.TYPE) + tasks }
|
override val tasks: Map<Name, Task<*>> by lazy { context.gather<Task<*>>(Task.TYPE) + tasks }
|
||||||
|
|
||||||
override suspend fun produce(taskName: Name, taskMeta: Meta): TaskResult<*> {
|
override suspend fun produce(taskName: Name, taskMeta: Meta): TaskResult<*> {
|
||||||
|
@ -4,13 +4,14 @@ import space.kscience.dataforge.data.Data
|
|||||||
import space.kscience.dataforge.data.await
|
import space.kscience.dataforge.data.await
|
||||||
import space.kscience.dataforge.io.*
|
import space.kscience.dataforge.io.*
|
||||||
import space.kscience.dataforge.misc.DFInternal
|
import space.kscience.dataforge.misc.DFInternal
|
||||||
|
import kotlin.reflect.typeOf
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Convert an [Envelope] to a data via given format. The actual parsing is done lazily.
|
* Convert an [Envelope] to a data via given format. The actual parsing is done lazily.
|
||||||
*/
|
*/
|
||||||
@OptIn(DFInternal::class)
|
@OptIn(DFInternal::class)
|
||||||
public fun <T : Any> Envelope.toData(format: IOReader<T>): Data<T> = Data(format.type, meta) {
|
public inline fun <reified T : Any> Envelope.toData(format: IOReader<T>): Data<T> = Data(typeOf<T>(), meta) {
|
||||||
data?.readWith(format) ?: error("Can't convert envelope without data to Data")
|
data?.readWith(format) ?: error("Can't convert envelope without data to Data")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2,9 +2,10 @@ package space.kscience.dataforge.workspace
|
|||||||
|
|
||||||
import space.kscience.dataforge.actions.Action
|
import space.kscience.dataforge.actions.Action
|
||||||
import space.kscience.dataforge.context.PluginFactory
|
import space.kscience.dataforge.context.PluginFactory
|
||||||
import space.kscience.dataforge.data.DataSet
|
import space.kscience.dataforge.data.DataTree
|
||||||
|
import space.kscience.dataforge.data.emitAll
|
||||||
import space.kscience.dataforge.data.forEach
|
import space.kscience.dataforge.data.forEach
|
||||||
import space.kscience.dataforge.data.map
|
import space.kscience.dataforge.data.transform
|
||||||
import space.kscience.dataforge.meta.*
|
import space.kscience.dataforge.meta.*
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
import space.kscience.dataforge.misc.DFExperimental
|
||||||
import space.kscience.dataforge.names.Name
|
import space.kscience.dataforge.names.Name
|
||||||
@ -27,13 +28,13 @@ public val TaskResultBuilder<*>.defaultDependencyMeta: Meta
|
|||||||
public suspend fun <T : Any> TaskResultBuilder<*>.from(
|
public suspend fun <T : Any> TaskResultBuilder<*>.from(
|
||||||
selector: DataSelector<T>,
|
selector: DataSelector<T>,
|
||||||
dependencyMeta: Meta = defaultDependencyMeta,
|
dependencyMeta: Meta = defaultDependencyMeta,
|
||||||
): DataSet<T> = selector.select(workspace, dependencyMeta)
|
): DataTree<T> = selector.select(workspace, dependencyMeta)
|
||||||
|
|
||||||
public suspend inline fun <T : Any, reified P : WorkspacePlugin> TaskResultBuilder<*>.from(
|
public suspend inline fun <T : Any, reified P : WorkspacePlugin> TaskResultBuilder<*>.from(
|
||||||
plugin: P,
|
plugin: P,
|
||||||
dependencyMeta: Meta = defaultDependencyMeta,
|
dependencyMeta: Meta = defaultDependencyMeta,
|
||||||
selectorBuilder: P.() -> TaskReference<T>,
|
selectorBuilder: P.() -> TaskReference<T>,
|
||||||
): DataSet<T> {
|
): TaskResult<T> {
|
||||||
require(workspace.context.plugins.contains(plugin)) { "Plugin $plugin is not loaded into $workspace" }
|
require(workspace.context.plugins.contains(plugin)) { "Plugin $plugin is not loaded into $workspace" }
|
||||||
val taskReference: TaskReference<T> = plugin.selectorBuilder()
|
val taskReference: TaskReference<T> = plugin.selectorBuilder()
|
||||||
val res = workspace.produce(plugin.name + taskReference.taskName, dependencyMeta)
|
val res = workspace.produce(plugin.name + taskReference.taskName, dependencyMeta)
|
||||||
@ -53,7 +54,7 @@ public suspend inline fun <reified T : Any, reified P : WorkspacePlugin> TaskRes
|
|||||||
pluginFactory: PluginFactory<P>,
|
pluginFactory: PluginFactory<P>,
|
||||||
dependencyMeta: Meta = defaultDependencyMeta,
|
dependencyMeta: Meta = defaultDependencyMeta,
|
||||||
selectorBuilder: P.() -> TaskReference<T>,
|
selectorBuilder: P.() -> TaskReference<T>,
|
||||||
): DataSet<T> {
|
): TaskResult<T> {
|
||||||
val plugin = workspace.context.plugins[pluginFactory]
|
val plugin = workspace.context.plugins[pluginFactory]
|
||||||
?: error("Plugin ${pluginFactory.tag} not loaded into workspace context")
|
?: error("Plugin ${pluginFactory.tag} not loaded into workspace context")
|
||||||
val taskReference: TaskReference<T> = plugin.selectorBuilder()
|
val taskReference: TaskReference<T> = plugin.selectorBuilder()
|
||||||
@ -64,9 +65,7 @@ public suspend inline fun <reified T : Any, reified P : WorkspacePlugin> TaskRes
|
|||||||
}
|
}
|
||||||
|
|
||||||
public val TaskResultBuilder<*>.allData: DataSelector<*>
|
public val TaskResultBuilder<*>.allData: DataSelector<*>
|
||||||
get() = object : DataSelector<Any> {
|
get() = DataSelector { workspace, _ -> workspace.data }
|
||||||
override suspend fun select(workspace: Workspace, meta: Meta): DataSet<Any> = workspace.data
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Perform a lazy mapping task using given [selector] and one-to-one [action].
|
* Perform a lazy mapping task using given [selector] and one-to-one [action].
|
||||||
@ -90,19 +89,19 @@ public suspend inline fun <T : Any, reified R : Any> TaskResultBuilder<R>.transf
|
|||||||
dataMetaTransform(data.name)
|
dataMetaTransform(data.name)
|
||||||
}
|
}
|
||||||
|
|
||||||
val res = data.map(meta, workspace.context.coroutineContext) {
|
val res = data.transform(meta, workspace.context.coroutineContext) {
|
||||||
action(it, data.name, meta)
|
action(it, data.name, meta)
|
||||||
}
|
}
|
||||||
|
|
||||||
data(data.name, res)
|
emit(data.name, res)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Set given [dataSet] as a task result.
|
* Set given [dataSet] as a task result.
|
||||||
*/
|
*/
|
||||||
public fun <T : Any> TaskResultBuilder<T>.result(dataSet: DataSet<T>) {
|
public fun <T : Any> TaskResultBuilder<T>.result(dataSet: DataTree<T>) {
|
||||||
node(Name.EMPTY, dataSet)
|
emitAll(dataSet)
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -114,7 +113,7 @@ public suspend inline fun <T : Any, reified R : Any> TaskResultBuilder<R>.action
|
|||||||
action: Action<T, R>,
|
action: Action<T, R>,
|
||||||
dependencyMeta: Meta = defaultDependencyMeta,
|
dependencyMeta: Meta = defaultDependencyMeta,
|
||||||
) {
|
) {
|
||||||
node(Name.EMPTY, action.execute(from(selector,dependencyMeta), dependencyMeta))
|
emitAll(action.execute(workspace.context, from(selector, dependencyMeta), dependencyMeta))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -1,5 +1,6 @@
|
|||||||
package space.kscience.dataforge.workspace
|
package space.kscience.dataforge.workspace
|
||||||
|
|
||||||
|
import kotlinx.coroutines.flow.map
|
||||||
import kotlinx.io.*
|
import kotlinx.io.*
|
||||||
import kotlinx.serialization.ExperimentalSerializationApi
|
import kotlinx.serialization.ExperimentalSerializationApi
|
||||||
import kotlinx.serialization.KSerializer
|
import kotlinx.serialization.KSerializer
|
||||||
@ -9,12 +10,10 @@ import kotlinx.serialization.serializer
|
|||||||
import space.kscience.dataforge.context.error
|
import space.kscience.dataforge.context.error
|
||||||
import space.kscience.dataforge.context.logger
|
import space.kscience.dataforge.context.logger
|
||||||
import space.kscience.dataforge.context.request
|
import space.kscience.dataforge.context.request
|
||||||
import space.kscience.dataforge.data.Data
|
import space.kscience.dataforge.data.*
|
||||||
import space.kscience.dataforge.data.await
|
|
||||||
import space.kscience.dataforge.io.*
|
import space.kscience.dataforge.io.*
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
import space.kscience.dataforge.misc.DFExperimental
|
||||||
import space.kscience.dataforge.misc.DFInternal
|
import space.kscience.dataforge.misc.DFInternal
|
||||||
import space.kscience.dataforge.names.Name
|
|
||||||
import space.kscience.dataforge.names.withIndex
|
import space.kscience.dataforge.names.withIndex
|
||||||
import java.nio.file.Path
|
import java.nio.file.Path
|
||||||
import kotlin.io.path.deleteIfExists
|
import kotlin.io.path.deleteIfExists
|
||||||
@ -22,7 +21,7 @@ import kotlin.io.path.div
|
|||||||
import kotlin.io.path.exists
|
import kotlin.io.path.exists
|
||||||
import kotlin.reflect.KType
|
import kotlin.reflect.KType
|
||||||
|
|
||||||
public class JsonIOFormat<T : Any>(override val type: KType) : IOFormat<T> {
|
public class JsonIOFormat<T>(private val type: KType) : IOFormat<T> {
|
||||||
|
|
||||||
@Suppress("UNCHECKED_CAST")
|
@Suppress("UNCHECKED_CAST")
|
||||||
private val serializer: KSerializer<T> = serializer(type) as KSerializer<T>
|
private val serializer: KSerializer<T> = serializer(type) as KSerializer<T>
|
||||||
@ -35,7 +34,7 @@ public class JsonIOFormat<T : Any>(override val type: KType) : IOFormat<T> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@OptIn(ExperimentalSerializationApi::class)
|
@OptIn(ExperimentalSerializationApi::class)
|
||||||
public class ProtobufIOFormat<T : Any>(override val type: KType) : IOFormat<T> {
|
public class ProtobufIOFormat<T>(private val type: KType) : IOFormat<T> {
|
||||||
|
|
||||||
@Suppress("UNCHECKED_CAST")
|
@Suppress("UNCHECKED_CAST")
|
||||||
private val serializer: KSerializer<T> = serializer(type) as KSerializer<T>
|
private val serializer: KSerializer<T> = serializer(type) as KSerializer<T>
|
||||||
@ -53,14 +52,14 @@ public class FileWorkspaceCache(public val cacheDirectory: Path) : WorkspaceCach
|
|||||||
// private fun <T : Any> TaskData<*>.checkType(taskType: KType): TaskData<T> = this as TaskData<T>
|
// private fun <T : Any> TaskData<*>.checkType(taskType: KType): TaskData<T> = this as TaskData<T>
|
||||||
|
|
||||||
@OptIn(DFExperimental::class, DFInternal::class)
|
@OptIn(DFExperimental::class, DFInternal::class)
|
||||||
override suspend fun <T : Any> evaluate(result: TaskResult<T>): TaskResult<T> {
|
override suspend fun <T> cache(result: TaskResult<T>): TaskResult<T> {
|
||||||
val io = result.workspace.context.request(IOPlugin)
|
val io = result.workspace.context.request(IOPlugin)
|
||||||
|
|
||||||
val format: IOFormat<T> = io.resolveIOFormat(result.dataType, result.taskMeta)
|
val format: IOFormat<T> = io.resolveIOFormat(result.dataType, result.taskMeta)
|
||||||
?: ProtobufIOFormat(result.dataType)
|
?: ProtobufIOFormat(result.dataType)
|
||||||
?: error("Can't resolve IOFormat for ${result.dataType}")
|
?: error("Can't resolve IOFormat for ${result.dataType}")
|
||||||
|
|
||||||
fun evaluateDatum(data: TaskData<T>): TaskData<T> {
|
fun cacheOne(data: NamedData<T>): NamedData<T> {
|
||||||
|
|
||||||
val path = cacheDirectory /
|
val path = cacheDirectory /
|
||||||
result.taskName.withIndex(result.taskMeta.hashCode().toString(16)).toString() /
|
result.taskName.withIndex(result.taskMeta.hashCode().toString(16)).toString() /
|
||||||
@ -92,15 +91,14 @@ public class FileWorkspaceCache(public val cacheDirectory: Path) : WorkspaceCach
|
|||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
return data.workspace.wrapData(datum, data.name, data.taskName, data.taskMeta)
|
return datum.named(data.name)
|
||||||
}
|
}
|
||||||
|
|
||||||
return object : TaskResult<T> by result {
|
|
||||||
override fun iterator(): Iterator<TaskData<T>> =
|
|
||||||
result.iterator().asSequence().map { evaluateDatum(it) }.iterator()
|
|
||||||
|
|
||||||
override fun get(name: Name): TaskData<T>? = result[name]?.let { evaluateDatum(it) }
|
val cachedTree = result.data.asSequence().map { cacheOne(it) }
|
||||||
}
|
.toObservableTree(result.dataType, result.workspace, result.data.updates().map { cacheOne(it) })
|
||||||
|
|
||||||
|
return result.workspace.wrapResult(cachedTree, result.taskName, result.taskMeta)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -18,7 +18,7 @@ public class InMemoryWorkspaceCache : WorkspaceCache {
|
|||||||
if (type.isSubtypeOf(taskType)) this as TaskData<T>
|
if (type.isSubtypeOf(taskType)) this as TaskData<T>
|
||||||
else error("Cached data type mismatch: expected $taskType but got $type")
|
else error("Cached data type mismatch: expected $taskType but got $type")
|
||||||
|
|
||||||
override suspend fun <T : Any> evaluate(result: TaskResult<T>): TaskResult<T> {
|
override suspend fun <T : Any> cache(result: TaskResult<T>): TaskResult<T> {
|
||||||
for (d: TaskData<T> in result) {
|
for (d: TaskData<T> in result) {
|
||||||
cache.getOrPut(result.taskName to result.taskMeta) { HashMap() }.getOrPut(d.name) { d }
|
cache.getOrPut(result.taskName to result.taskMeta) { HashMap() }.getOrPut(d.name) { d }
|
||||||
}
|
}
|
||||||
|
@ -14,11 +14,8 @@ import space.kscience.dataforge.meta.copy
|
|||||||
import space.kscience.dataforge.meta.string
|
import space.kscience.dataforge.meta.string
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
import space.kscience.dataforge.misc.DFExperimental
|
||||||
import space.kscience.dataforge.misc.DFInternal
|
import space.kscience.dataforge.misc.DFInternal
|
||||||
import space.kscience.dataforge.names.Name
|
import space.kscience.dataforge.names.*
|
||||||
import space.kscience.dataforge.names.NameToken
|
import space.kscience.dataforge.workspace.FileData.Companion.defaultPathToName
|
||||||
import space.kscience.dataforge.names.asName
|
|
||||||
import space.kscience.dataforge.names.plus
|
|
||||||
import space.kscience.dataforge.workspace.FileData.Companion.DEFAULT_IGNORE_EXTENSIONS
|
|
||||||
import java.nio.file.Files
|
import java.nio.file.Files
|
||||||
import java.nio.file.Path
|
import java.nio.file.Path
|
||||||
import java.nio.file.StandardWatchEventKinds
|
import java.nio.file.StandardWatchEventKinds
|
||||||
@ -35,6 +32,7 @@ import kotlin.reflect.typeOf
|
|||||||
|
|
||||||
public typealias FileFormatResolver<T> = (path: Path, meta: Meta) -> IOReader<T>?
|
public typealias FileFormatResolver<T> = (path: Path, meta: Meta) -> IOReader<T>?
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A data based on a filesystem [Path]
|
* A data based on a filesystem [Path]
|
||||||
*/
|
*/
|
||||||
@ -54,13 +52,28 @@ public class FileData<T> internal constructor(private val data: Data<T>, public
|
|||||||
public val FILE_UPDATE_TIME_KEY: Name = FILE_KEY + "updated"
|
public val FILE_UPDATE_TIME_KEY: Name = FILE_KEY + "updated"
|
||||||
public const val DF_FILE_EXTENSION: String = "df"
|
public const val DF_FILE_EXTENSION: String = "df"
|
||||||
public val DEFAULT_IGNORE_EXTENSIONS: Set<String> = setOf(DF_FILE_EXTENSION)
|
public val DEFAULT_IGNORE_EXTENSIONS: Set<String> = setOf(DF_FILE_EXTENSION)
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Transform file name into DataForg name. Ignores DataForge file extensions.
|
||||||
|
*/
|
||||||
|
public val defaultPathToName: (Path) -> Name = { path ->
|
||||||
|
Name(
|
||||||
|
path.map { segment ->
|
||||||
|
if (segment.isRegularFile() && segment.extension in DEFAULT_IGNORE_EXTENSIONS) {
|
||||||
|
NameToken(path.nameWithoutExtension)
|
||||||
|
} else {
|
||||||
|
NameToken(path.name)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Read data with supported envelope format and binary format. If envelope format is null, then read binary directly from file.
|
* Read data with supported envelope format and binary format. If the envelope format is null, then read binary directly from file.
|
||||||
* The operation is blocking since it must read meta header. The reading of envelope body is lazy
|
* The operation is blocking since it must read the meta header. The reading of envelope body is lazy
|
||||||
*/
|
*/
|
||||||
@OptIn(DFInternal::class)
|
@OptIn(DFInternal::class)
|
||||||
@DFExperimental
|
@DFExperimental
|
||||||
@ -90,7 +103,7 @@ public fun <T : Any> IOPlugin.readDataFile(
|
|||||||
context(IOPlugin) @DFExperimental
|
context(IOPlugin) @DFExperimental
|
||||||
public fun <T : Any> DataSetBuilder<T>.directory(
|
public fun <T : Any> DataSetBuilder<T>.directory(
|
||||||
path: Path,
|
path: Path,
|
||||||
ignoreExtensions: Set<String>,
|
pathToName: (Path) -> Name = defaultPathToName,
|
||||||
formatResolver: FileFormatResolver<T>,
|
formatResolver: FileFormatResolver<T>,
|
||||||
) {
|
) {
|
||||||
Files.list(path).forEach { childPath ->
|
Files.list(path).forEach { childPath ->
|
||||||
@ -98,7 +111,7 @@ public fun <T : Any> DataSetBuilder<T>.directory(
|
|||||||
if (fileName.startsWith(IOPlugin.META_FILE_NAME)) {
|
if (fileName.startsWith(IOPlugin.META_FILE_NAME)) {
|
||||||
meta(readMetaFile(childPath))
|
meta(readMetaFile(childPath))
|
||||||
} else if (!fileName.startsWith("@")) {
|
} else if (!fileName.startsWith("@")) {
|
||||||
file(childPath, ignoreExtensions, formatResolver)
|
file(childPath, pathToName, formatResolver)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -111,9 +124,9 @@ public fun <T : Any> DataSetBuilder<T>.directory(
|
|||||||
public fun <T : Any> IOPlugin.readDataDirectory(
|
public fun <T : Any> IOPlugin.readDataDirectory(
|
||||||
type: KType,
|
type: KType,
|
||||||
path: Path,
|
path: Path,
|
||||||
ignoreExtensions: Set<String> = DEFAULT_IGNORE_EXTENSIONS,
|
pathToName: (Path) -> Name = defaultPathToName,
|
||||||
formatResolver: FileFormatResolver<T>,
|
formatResolver: FileFormatResolver<T>,
|
||||||
): DataTree<T> {
|
): LegacyDataTree<T> {
|
||||||
//read zipped data node
|
//read zipped data node
|
||||||
if (path.fileName != null && path.fileName.toString().endsWith(".zip")) {
|
if (path.fileName != null && path.fileName.toString().endsWith(".zip")) {
|
||||||
//Using explicit Zip file system to avoid bizarre compatibility bugs
|
//Using explicit Zip file system to avoid bizarre compatibility bugs
|
||||||
@ -121,14 +134,14 @@ public fun <T : Any> IOPlugin.readDataDirectory(
|
|||||||
?: error("Zip file system provider not found")
|
?: error("Zip file system provider not found")
|
||||||
val fs = fsProvider.newFileSystem(path, mapOf("create" to "true"))
|
val fs = fsProvider.newFileSystem(path, mapOf("create" to "true"))
|
||||||
|
|
||||||
return readDataDirectory(type, fs.rootDirectories.first(), ignoreExtensions, formatResolver)
|
return readDataDirectory(type, fs.rootDirectories.first(), pathToName, formatResolver)
|
||||||
}
|
}
|
||||||
if (!Files.isDirectory(path)) error("Provided path $path is not a directory")
|
if (!Files.isDirectory(path)) error("Provided path $path is not a directory")
|
||||||
return DataTree(type) {
|
return DataTree(type) {
|
||||||
meta {
|
meta {
|
||||||
FileData.FILE_PATH_KEY put path.toString()
|
FileData.FILE_PATH_KEY put path.toString()
|
||||||
}
|
}
|
||||||
directory(path, ignoreExtensions, formatResolver)
|
directory(path, pathToName, formatResolver)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -136,9 +149,9 @@ public fun <T : Any> IOPlugin.readDataDirectory(
|
|||||||
@DFExperimental
|
@DFExperimental
|
||||||
public inline fun <reified T : Any> IOPlugin.readDataDirectory(
|
public inline fun <reified T : Any> IOPlugin.readDataDirectory(
|
||||||
path: Path,
|
path: Path,
|
||||||
ignoreExtensions: Set<String> = DEFAULT_IGNORE_EXTENSIONS,
|
noinline pathToName: (Path) -> Name = defaultPathToName,
|
||||||
noinline formatResolver: FileFormatResolver<T>,
|
noinline formatResolver: FileFormatResolver<T>,
|
||||||
): DataTree<T> = readDataDirectory(typeOf<T>(), path, ignoreExtensions, formatResolver)
|
): LegacyDataTree<T> = readDataDirectory(typeOf<T>(), path, pathToName, formatResolver)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Read a raw binary data tree from the directory. All files are read as-is (save for meta files).
|
* Read a raw binary data tree from the directory. All files are read as-is (save for meta files).
|
||||||
@ -146,8 +159,8 @@ public inline fun <reified T : Any> IOPlugin.readDataDirectory(
|
|||||||
@DFExperimental
|
@DFExperimental
|
||||||
public fun IOPlugin.readRawDirectory(
|
public fun IOPlugin.readRawDirectory(
|
||||||
path: Path,
|
path: Path,
|
||||||
ignoreExtensions: Set<String> = emptySet(),
|
pathToName: (Path) -> Name = defaultPathToName,
|
||||||
): DataTree<Binary> = readDataDirectory(path, ignoreExtensions) { _, _ -> IOReader.binary }
|
): LegacyDataTree<Binary> = readDataDirectory(path, pathToName) { _, _ -> IOReader.binary }
|
||||||
|
|
||||||
|
|
||||||
private fun Path.toName() = Name(map { NameToken.parse(it.nameWithoutExtension) })
|
private fun Path.toName() = Name(map { NameToken.parse(it.nameWithoutExtension) })
|
||||||
@ -157,13 +170,13 @@ private fun Path.toName() = Name(map { NameToken.parse(it.nameWithoutExtension)
|
|||||||
public fun <T : Any> IOPlugin.monitorDataDirectory(
|
public fun <T : Any> IOPlugin.monitorDataDirectory(
|
||||||
type: KType,
|
type: KType,
|
||||||
path: Path,
|
path: Path,
|
||||||
ignoreExtensions: Set<String> = DEFAULT_IGNORE_EXTENSIONS,
|
pathToName: (Path) -> Name = defaultPathToName,
|
||||||
formatResolver: FileFormatResolver<T>,
|
formatResolver: FileFormatResolver<T>,
|
||||||
): DataSource<T> {
|
): DataSource<T> {
|
||||||
if (path.fileName.toString().endsWith(".zip")) error("Monitoring not supported for ZipFS")
|
if (path.fileName.toString().endsWith(".zip")) error("Monitoring not supported for ZipFS")
|
||||||
if (!Files.isDirectory(path)) error("Provided path $path is not a directory")
|
if (!Files.isDirectory(path)) error("Provided path $path is not a directory")
|
||||||
return DataSource(type, context) {
|
return DataSource(type, context) {
|
||||||
directory(path, ignoreExtensions, formatResolver)
|
directory(path, pathToName, formatResolver)
|
||||||
launch(Dispatchers.IO) {
|
launch(Dispatchers.IO) {
|
||||||
val watchService = path.fileSystem.newWatchService()
|
val watchService = path.fileSystem.newWatchService()
|
||||||
|
|
||||||
@ -186,7 +199,7 @@ public fun <T : Any> IOPlugin.monitorDataDirectory(
|
|||||||
if (fileName.startsWith(IOPlugin.META_FILE_NAME)) {
|
if (fileName.startsWith(IOPlugin.META_FILE_NAME)) {
|
||||||
meta(readMetaFile(eventPath))
|
meta(readMetaFile(eventPath))
|
||||||
} else if (!fileName.startsWith("@")) {
|
} else if (!fileName.startsWith("@")) {
|
||||||
file(eventPath, ignoreExtensions, formatResolver)
|
file(eventPath, pathToName, formatResolver)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -205,9 +218,9 @@ public fun <T : Any> IOPlugin.monitorDataDirectory(
|
|||||||
@DFExperimental
|
@DFExperimental
|
||||||
public inline fun <reified T : Any> IOPlugin.monitorDataDirectory(
|
public inline fun <reified T : Any> IOPlugin.monitorDataDirectory(
|
||||||
path: Path,
|
path: Path,
|
||||||
ignoreExtensions: Set<String> = DEFAULT_IGNORE_EXTENSIONS,
|
noinline pathToName: (Path) -> Name = defaultPathToName,
|
||||||
noinline formatResolver: FileFormatResolver<T>,
|
noinline formatResolver: FileFormatResolver<T>,
|
||||||
): DataSource<T> = monitorDataDirectory(typeOf<T>(), path, ignoreExtensions, formatResolver)
|
): DataSource<T> = monitorDataDirectory(typeOf<T>(), path, pathToName, formatResolver)
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Read and monitor raw binary data tree from the directory. All files are read as-is (save for meta files).
|
* Read and monitor raw binary data tree from the directory. All files are read as-is (save for meta files).
|
||||||
@ -215,18 +228,23 @@ public inline fun <reified T : Any> IOPlugin.monitorDataDirectory(
|
|||||||
@DFExperimental
|
@DFExperimental
|
||||||
public fun IOPlugin.monitorRawDirectory(
|
public fun IOPlugin.monitorRawDirectory(
|
||||||
path: Path,
|
path: Path,
|
||||||
ignoreExtensions: Set<String> = DEFAULT_IGNORE_EXTENSIONS,
|
pathToName: (Path) -> Name = defaultPathToName,
|
||||||
): DataSource<Binary> = monitorDataDirectory(path, ignoreExtensions) { _, _ -> IOReader.binary }
|
): DataSource<Binary> = monitorDataDirectory(path, pathToName) { _, _ -> IOReader.binary }
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Write data tree to existing directory or create a new one using default [java.nio.file.FileSystem] provider
|
* Write the data tree to existing directory or create a new one using default [java.nio.file.FileSystem] provider
|
||||||
|
*
|
||||||
|
* @param nameToPath a [Name] to [Path] converter used to create
|
||||||
*/
|
*/
|
||||||
@DFExperimental
|
@DFExperimental
|
||||||
public suspend fun <T : Any> IOPlugin.writeDataDirectory(
|
public suspend fun <T : Any> IOPlugin.writeDataDirectory(
|
||||||
path: Path,
|
path: Path,
|
||||||
tree: DataTree<T>,
|
dataSet: DataSet<T>,
|
||||||
format: IOWriter<T>,
|
format: IOWriter<T>,
|
||||||
envelopeFormat: EnvelopeFormat? = null,
|
envelopeFormat: EnvelopeFormat? = null,
|
||||||
|
nameToPath: (name: Name, data: Data<T>) -> Path = { name, _ ->
|
||||||
|
Path(name.tokens.joinToString("/") { token -> token.toStringUnescaped() })
|
||||||
|
},
|
||||||
) {
|
) {
|
||||||
withContext(Dispatchers.IO) {
|
withContext(Dispatchers.IO) {
|
||||||
if (!Files.exists(path)) {
|
if (!Files.exists(path)) {
|
||||||
@ -234,67 +252,54 @@ public suspend fun <T : Any> IOPlugin.writeDataDirectory(
|
|||||||
} else if (!Files.isDirectory(path)) {
|
} else if (!Files.isDirectory(path)) {
|
||||||
error("Can't write a node into file")
|
error("Can't write a node into file")
|
||||||
}
|
}
|
||||||
tree.items.forEach { (token, item) ->
|
dataSet.forEach { (name, data) ->
|
||||||
val childPath = path.resolve(token.toString())
|
val childPath = path.resolve(nameToPath(name, data))
|
||||||
when (item) {
|
childPath.parent.createDirectories()
|
||||||
is DataTreeItem.Node -> {
|
val envelope = data.toEnvelope(format)
|
||||||
writeDataDirectory(childPath, item.tree, format, envelopeFormat)
|
|
||||||
}
|
|
||||||
|
|
||||||
is DataTreeItem.Leaf -> {
|
|
||||||
val envelope = item.data.toEnvelope(format)
|
|
||||||
if (envelopeFormat != null) {
|
if (envelopeFormat != null) {
|
||||||
writeEnvelopeFile(childPath, envelope, envelopeFormat)
|
writeEnvelopeFile(childPath, envelope, envelopeFormat)
|
||||||
} else {
|
} else {
|
||||||
writeEnvelopeDirectory(childPath, envelope)
|
writeEnvelopeDirectory(childPath, envelope)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
val directoryMeta = dataSet.meta
|
||||||
}
|
writeMetaFile(path, directoryMeta)
|
||||||
val treeMeta = tree.meta
|
|
||||||
writeMetaFile(path, treeMeta)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Reads the specified resources and returns a [DataTree] containing the data.
|
* Reads the specified resources and returns a [LegacyDataTree] containing the data.
|
||||||
*
|
*
|
||||||
* @param resources The names of the resources to read.
|
* @param resources The names of the resources to read.
|
||||||
* @param classLoader The class loader to use for loading the resources. By default, it uses the current thread's context class loader.
|
* @param classLoader The class loader to use for loading the resources. By default, it uses the current thread's context class loader.
|
||||||
* @return A DataTree containing the data read from the resources.
|
* @return A DataTree containing the data read from the resources.
|
||||||
*/
|
*/
|
||||||
@DFExperimental
|
@DFExperimental
|
||||||
private fun IOPlugin.readResources(
|
public fun IOPlugin.readResources(
|
||||||
vararg resources: String,
|
vararg resources: String,
|
||||||
|
pathToName: (Path) -> Name = defaultPathToName,
|
||||||
classLoader: ClassLoader = Thread.currentThread().contextClassLoader,
|
classLoader: ClassLoader = Thread.currentThread().contextClassLoader,
|
||||||
): DataTree<Binary> {
|
): LegacyDataTree<Binary> = GenericDataTree {
|
||||||
// require(resource.isNotBlank()) {"Can't mount root resource tree as data root"}
|
|
||||||
return DataTree {
|
|
||||||
resources.forEach { resource ->
|
resources.forEach { resource ->
|
||||||
val path = classLoader.getResource(resource)?.toURI()?.toPath() ?: error(
|
val path = classLoader.getResource(resource)?.toURI()?.toPath() ?: error(
|
||||||
"Resource with name $resource is not resolved"
|
"Resource with name $resource is not resolved"
|
||||||
)
|
)
|
||||||
node(resource, readRawDirectory(path))
|
node(resource, readRawDirectory(path, pathToName))
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Add file/directory-based data tree item
|
* Add file/directory-based data tree item
|
||||||
*
|
|
||||||
* @param ignoreExtensions a list of file extensions for which extension should be cut from the resulting item name
|
|
||||||
*/
|
*/
|
||||||
context(IOPlugin)
|
context(IOPlugin)
|
||||||
@OptIn(DFInternal::class)
|
@OptIn(DFInternal::class)
|
||||||
@DFExperimental
|
@DFExperimental
|
||||||
public fun <T : Any> DataSetBuilder<T>.file(
|
public fun <T : Any> DataSetBuilder<T>.file(
|
||||||
path: Path,
|
path: Path,
|
||||||
ignoreExtensions: Set<String> = DEFAULT_IGNORE_EXTENSIONS,
|
pathToName: (Path) -> Name = defaultPathToName,
|
||||||
formatResolver: FileFormatResolver<out T>,
|
formatResolver: FileFormatResolver<out T>,
|
||||||
) {
|
) {
|
||||||
|
|
||||||
fun defaultPath() = if (path.extension in ignoreExtensions) path.nameWithoutExtension else path.name
|
|
||||||
|
|
||||||
try {
|
try {
|
||||||
//If path is a single file or a special directory, read it as single datum
|
//If path is a single file or a special directory, read it as single datum
|
||||||
if (!Files.isDirectory(path) || Files.list(path).allMatch { it.fileName.toString().startsWith("@") }) {
|
if (!Files.isDirectory(path) || Files.list(path).allMatch { it.fileName.toString().startsWith("@") }) {
|
||||||
@ -303,16 +308,16 @@ public fun <T : Any> DataSetBuilder<T>.file(
|
|||||||
logger.warn { "File format is not resolved for $path. Skipping." }
|
logger.warn { "File format is not resolved for $path. Skipping." }
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
val name: String = data.meta[Envelope.ENVELOPE_NAME_KEY].string ?: defaultPath()
|
val name: Name = data.meta[Envelope.ENVELOPE_NAME_KEY].string?.parseAsName() ?: pathToName(path.last())
|
||||||
data(name.asName(), data)
|
data(name, data)
|
||||||
} else {
|
} else {
|
||||||
//otherwise, read as directory
|
//otherwise, read as directory
|
||||||
val data: DataTree<T> = readDataDirectory(dataType, path, ignoreExtensions, formatResolver)
|
val data: LegacyDataTree<T> = readDataDirectory(dataType, path, pathToName, formatResolver)
|
||||||
val name = data.meta[Envelope.ENVELOPE_NAME_KEY].string ?: defaultPath()
|
val name = data.meta[Envelope.ENVELOPE_NAME_KEY].string?.parseAsName() ?: pathToName(path.last())
|
||||||
node(name.asName(), data)
|
node(name, data)
|
||||||
}
|
}
|
||||||
} catch (ex: Exception) {
|
} catch (ex: Exception) {
|
||||||
logger.error { "Failed to read file or directory at $path: ${ex.message}" }
|
logger.error(ex) { "Failed to read file or directory at $path: ${ex.message}" }
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
package space.kscience.dataforge.workspace
|
package space.kscience.dataforge.workspace
|
||||||
|
|
||||||
import space.kscience.dataforge.data.DataSet
|
import space.kscience.dataforge.data.DataTree
|
||||||
import space.kscience.dataforge.data.filterByType
|
import space.kscience.dataforge.data.filterByType
|
||||||
import space.kscience.dataforge.meta.Meta
|
import space.kscience.dataforge.meta.Meta
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
import space.kscience.dataforge.misc.DFExperimental
|
||||||
@ -16,8 +16,7 @@ import space.kscience.dataforge.names.matches
|
|||||||
*/
|
*/
|
||||||
@OptIn(DFExperimental::class)
|
@OptIn(DFExperimental::class)
|
||||||
public inline fun <reified T : Any> TaskResultBuilder<*>.dataByType(namePattern: Name? = null): DataSelector<T> =
|
public inline fun <reified T : Any> TaskResultBuilder<*>.dataByType(namePattern: Name? = null): DataSelector<T> =
|
||||||
object : DataSelector<T> {
|
DataSelector<T> { workspace, meta ->
|
||||||
override suspend fun select(workspace: Workspace, meta: Meta): DataSet<T> =
|
|
||||||
workspace.data.filterByType { name, _ ->
|
workspace.data.filterByType { name, _ ->
|
||||||
namePattern == null || name.matches(namePattern)
|
namePattern == null || name.matches(namePattern)
|
||||||
}
|
}
|
||||||
@ -26,4 +25,4 @@ public inline fun <reified T : Any> TaskResultBuilder<*>.dataByType(namePattern:
|
|||||||
public suspend inline fun <reified T : Any> TaskResultBuilder<*>.fromTask(
|
public suspend inline fun <reified T : Any> TaskResultBuilder<*>.fromTask(
|
||||||
task: Name,
|
task: Name,
|
||||||
taskMeta: Meta = Meta.EMPTY,
|
taskMeta: Meta = Meta.EMPTY,
|
||||||
): DataSet<T> = workspace.produce(task, taskMeta).filterByType()
|
): DataTree<T> = workspace.produce(task, taskMeta).data.filterByType()
|
@ -2,8 +2,8 @@ package space.kscience.dataforge.workspace
|
|||||||
|
|
||||||
import kotlinx.coroutines.Dispatchers
|
import kotlinx.coroutines.Dispatchers
|
||||||
import kotlinx.coroutines.withContext
|
import kotlinx.coroutines.withContext
|
||||||
import space.kscience.dataforge.data.DataTree
|
|
||||||
import space.kscience.dataforge.data.DataTreeItem
|
import space.kscience.dataforge.data.DataTreeItem
|
||||||
|
import space.kscience.dataforge.data.LegacyDataTree
|
||||||
import space.kscience.dataforge.io.*
|
import space.kscience.dataforge.io.*
|
||||||
import space.kscience.dataforge.misc.DFExperimental
|
import space.kscience.dataforge.misc.DFExperimental
|
||||||
import java.nio.file.Files
|
import java.nio.file.Files
|
||||||
@ -47,10 +47,10 @@ private suspend fun <T : Any> ZipOutputStream.writeNode(
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Write this [DataTree] as a zip archive
|
* Write this [LegacyDataTree] as a zip archive
|
||||||
*/
|
*/
|
||||||
@DFExperimental
|
@DFExperimental
|
||||||
public suspend fun <T : Any> DataTree<T>.writeZip(
|
public suspend fun <T : Any> LegacyDataTree<T>.writeZip(
|
||||||
path: Path,
|
path: Path,
|
||||||
format: IOFormat<T>,
|
format: IOFormat<T>,
|
||||||
envelopeFormat: EnvelopeFormat = TaggedEnvelopeFormat,
|
envelopeFormat: EnvelopeFormat = TaggedEnvelopeFormat,
|
||||||
|
@ -20,13 +20,13 @@ class DataPropagationTestPlugin : WorkspacePlugin() {
|
|||||||
val result: Data<Int> = selectedData.foldToData(0) { result, data ->
|
val result: Data<Int> = selectedData.foldToData(0) { result, data ->
|
||||||
result + data.value
|
result + data.value
|
||||||
}
|
}
|
||||||
data("result", result)
|
emit("result", result)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
val singleData by task<Int> {
|
val singleData by task<Int> {
|
||||||
workspace.data.filterByType<Int>()["myData[12]"]?.let {
|
workspace.data.filterByType<Int>()["myData[12]"]?.let {
|
||||||
data("result", it)
|
emit("result", it)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -26,7 +26,7 @@ import kotlin.test.assertEquals
|
|||||||
|
|
||||||
|
|
||||||
class FileDataTest {
|
class FileDataTest {
|
||||||
val dataNode = DataTree<String> {
|
val dataNode = GenericDataTree<String> {
|
||||||
node("dir") {
|
node("dir") {
|
||||||
static("a", "Some string") {
|
static("a", "Some string") {
|
||||||
"content" put "Some string"
|
"content" put "Some string"
|
||||||
|
@ -140,7 +140,7 @@ internal class SimpleWorkspaceTest {
|
|||||||
val meta = data.meta.toMutableMeta().apply {
|
val meta = data.meta.toMutableMeta().apply {
|
||||||
"newValue" put 22
|
"newValue" put 22
|
||||||
}
|
}
|
||||||
data(data.name + "new", data.map { (data.meta["value"].int ?: 0) + it })
|
data(data.name + "new", data.transform { (data.meta["value"].int ?: 0) + it })
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user