Migration to new Storage API
This commit is contained in:
parent
32732c9de6
commit
da0b896d29
@ -31,7 +31,8 @@ import hep.dataforge.meta.Meta
|
|||||||
import hep.dataforge.states.StateDef
|
import hep.dataforge.states.StateDef
|
||||||
import hep.dataforge.states.valueState
|
import hep.dataforge.states.valueState
|
||||||
import hep.dataforge.storage.StorageConnection
|
import hep.dataforge.storage.StorageConnection
|
||||||
import hep.dataforge.storage.TableLoader
|
import hep.dataforge.storage.tables.TableLoader
|
||||||
|
import hep.dataforge.storage.tables.createTable
|
||||||
import hep.dataforge.tables.TableFormat
|
import hep.dataforge.tables.TableFormat
|
||||||
import hep.dataforge.tables.TableFormatBuilder
|
import hep.dataforge.tables.TableFormatBuilder
|
||||||
import hep.dataforge.utils.DateTimeUtils
|
import hep.dataforge.utils.DateTimeUtils
|
||||||
@ -91,13 +92,7 @@ class PKT8Device(context: Context, meta: Meta) : PortSensor(context, meta) {
|
|||||||
private fun buildLoader(connection: StorageConnection): TableLoader {
|
private fun buildLoader(connection: StorageConnection): TableLoader {
|
||||||
val storage = connection.storage
|
val storage = connection.storage
|
||||||
val suffix = DateTimeUtils.fileSuffix()
|
val suffix = DateTimeUtils.fileSuffix()
|
||||||
|
return storage.createTable("cryotemp_$suffix", tableFormat)
|
||||||
try {
|
|
||||||
return LoaderFactory.buildPointLoader(storage, "cryotemp_$suffix", "", "timestamp", tableFormat)
|
|
||||||
} catch (e: StorageException) {
|
|
||||||
throw RuntimeException("Failed to builder loader from storage", e)
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Throws(ControlException::class)
|
@Throws(ControlException::class)
|
||||||
|
@ -16,12 +16,19 @@
|
|||||||
|
|
||||||
package inr.numass.control.dante
|
package inr.numass.control.dante
|
||||||
|
|
||||||
|
import hep.dataforge.context.Context
|
||||||
|
import hep.dataforge.context.ContextAware
|
||||||
|
import hep.dataforge.context.launch
|
||||||
import hep.dataforge.meta.Meta
|
import hep.dataforge.meta.Meta
|
||||||
import hep.dataforge.meta.buildMeta
|
import hep.dataforge.meta.buildMeta
|
||||||
|
import hep.dataforge.orElse
|
||||||
import inr.numass.control.dante.DanteClient.Companion.CommandType.*
|
import inr.numass.control.dante.DanteClient.Companion.CommandType.*
|
||||||
import inr.numass.control.dante.DanteClient.Companion.Register.*
|
import inr.numass.control.dante.DanteClient.Companion.Register.*
|
||||||
|
import inr.numass.data.NumassProto
|
||||||
|
import inr.numass.data.ProtoNumassPoint
|
||||||
import inr.numass.data.api.NumassPoint
|
import inr.numass.data.api.NumassPoint
|
||||||
import inr.numass.data.storage.ProtoNumassPoint
|
import kotlinx.coroutines.*
|
||||||
|
import kotlinx.coroutines.channels.Channel
|
||||||
import java.io.DataInputStream
|
import java.io.DataInputStream
|
||||||
import java.io.OutputStream
|
import java.io.OutputStream
|
||||||
import java.lang.Math.pow
|
import java.lang.Math.pow
|
||||||
@ -64,15 +71,13 @@ internal val ByteArray.hex
|
|||||||
|
|
||||||
|
|
||||||
//TODO implement using Device
|
//TODO implement using Device
|
||||||
class DanteClient(val ip: String, chainLength: Int) : AutoCloseable {
|
class DanteClient(override val context: Context,val ip: String, chainLength: Int) : AutoCloseable, ContextAware {
|
||||||
private val RESET_COMMAND = byteArrayOf(0xDD.toByte(), 0x55, 0xDD.toByte(), 0xEE.toByte())
|
private val RESET_COMMAND = byteArrayOf(0xDD.toByte(), 0x55, 0xDD.toByte(), 0xEE.toByte())
|
||||||
|
|
||||||
private val logger = LoggerFactory.getLogger(javaClass)
|
|
||||||
|
|
||||||
private val packetNumber = AtomicLong(0)
|
private val packetNumber = AtomicLong(0)
|
||||||
|
|
||||||
private lateinit var parentJob: Job
|
private val parentJob: Job = SupervisorJob()
|
||||||
private val pool = newFixedThreadPoolContext(8, "Dante")
|
private val pool = newFixedThreadPoolContext(8, "Dante") + parentJob
|
||||||
|
|
||||||
private val connections: MutableMap<Int, Pair<Socket, Job>> = HashMap()
|
private val connections: MutableMap<Int, Pair<Socket, Job>> = HashMap()
|
||||||
|
|
||||||
@ -101,8 +106,8 @@ class DanteClient(val ip: String, chainLength: Int) : AutoCloseable {
|
|||||||
|
|
||||||
|
|
||||||
fun open() {
|
fun open() {
|
||||||
//create a new parent job
|
//start supervisor job
|
||||||
this.parentJob = Job()
|
parentJob.start()
|
||||||
(0..3).forEach {
|
(0..3).forEach {
|
||||||
openPort(it)
|
openPort(it)
|
||||||
}
|
}
|
||||||
@ -114,7 +119,7 @@ class DanteClient(val ip: String, chainLength: Int) : AutoCloseable {
|
|||||||
it.first.close()
|
it.first.close()
|
||||||
it.second.cancel()
|
it.second.cancel()
|
||||||
}
|
}
|
||||||
parentJob.cancel(CancellationException("Server stopped"))
|
parentJob.cancel()
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -132,7 +137,7 @@ class DanteClient(val ip: String, chainLength: Int) : AutoCloseable {
|
|||||||
* Reset everything
|
* Reset everything
|
||||||
*/
|
*/
|
||||||
fun reset() {
|
fun reset() {
|
||||||
async {
|
launch {
|
||||||
sendChannel.send(RESET_COMMAND)
|
sendChannel.send(RESET_COMMAND)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -152,7 +157,7 @@ class DanteClient(val ip: String, chainLength: Int) : AutoCloseable {
|
|||||||
if (port == 0) {
|
if (port == 0) {
|
||||||
//outputJob.cancel()
|
//outputJob.cancel()
|
||||||
output = socket.getOutputStream()
|
output = socket.getOutputStream()
|
||||||
outputJob = launch(context = pool, parent = parentJob) {
|
outputJob = launch(pool) {
|
||||||
while (true) {
|
while (true) {
|
||||||
val command = sendChannel.receive()
|
val command = sendChannel.receive()
|
||||||
output.write(command)
|
output.write(command)
|
||||||
@ -162,7 +167,7 @@ class DanteClient(val ip: String, chainLength: Int) : AutoCloseable {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
val job = launch(context = pool, parent = parentJob) {
|
val job = launch(pool) {
|
||||||
val stream = socket.getInputStream()
|
val stream = socket.getInputStream()
|
||||||
while (true) {
|
while (true) {
|
||||||
if (stream.read() == PACKET_HEADER_START_BYTES[0] && stream.read() == PACKET_HEADER_START_BYTES[1]) {
|
if (stream.read() == PACKET_HEADER_START_BYTES[0] && stream.read() == PACKET_HEADER_START_BYTES[1]) {
|
||||||
@ -249,7 +254,7 @@ class DanteClient(val ip: String, chainLength: Int) : AutoCloseable {
|
|||||||
}
|
}
|
||||||
|
|
||||||
return sequence {
|
return sequence {
|
||||||
val intBuffer = ByteBuffer.wrap(message.payload).asIntBuffer()
|
val intBuffer = ByteBuffer.wrap(message!!.payload).asIntBuffer()
|
||||||
while (intBuffer.hasRemaining()) {
|
while (intBuffer.hasRemaining()) {
|
||||||
yield(intBuffer.get())
|
yield(intBuffer.get())
|
||||||
}
|
}
|
||||||
|
@ -29,10 +29,11 @@ import inr.numass.data.analyzers.NumassAnalyzer
|
|||||||
import inr.numass.data.analyzers.SimpleAnalyzer
|
import inr.numass.data.analyzers.SimpleAnalyzer
|
||||||
import inr.numass.data.analyzers.withBinning
|
import inr.numass.data.analyzers.withBinning
|
||||||
import inr.numass.data.api.NumassBlock
|
import inr.numass.data.api.NumassBlock
|
||||||
import inr.numass.data.channel
|
import kotlinx.coroutines.runBlocking
|
||||||
|
|
||||||
|
|
||||||
fun main(args: Array<String>) {
|
fun main(args: Array<String>) {
|
||||||
val client = DanteClient("192.168.111.120", 8)
|
val client = DanteClient(Global,"192.168.111.120", 8)
|
||||||
client.open()
|
client.open()
|
||||||
val meta = buildMeta {
|
val meta = buildMeta {
|
||||||
"gain" to 1.0
|
"gain" to 1.0
|
||||||
|
@ -15,13 +15,23 @@
|
|||||||
*/
|
*/
|
||||||
package inr.numass.control.magnet
|
package inr.numass.control.magnet
|
||||||
|
|
||||||
|
import hep.dataforge.context.Context
|
||||||
import hep.dataforge.control.devices.AbstractDevice
|
import hep.dataforge.control.devices.AbstractDevice
|
||||||
import hep.dataforge.control.ports.Port
|
import hep.dataforge.control.ports.Port
|
||||||
import hep.dataforge.control.ports.PortFactory
|
import hep.dataforge.control.ports.PortFactory
|
||||||
|
import hep.dataforge.description.ValueDef
|
||||||
import hep.dataforge.exceptions.ControlException
|
import hep.dataforge.exceptions.ControlException
|
||||||
import hep.dataforge.exceptions.PortException
|
import hep.dataforge.exceptions.PortException
|
||||||
|
import hep.dataforge.meta.Meta
|
||||||
|
import hep.dataforge.meta.buildMeta
|
||||||
|
import hep.dataforge.states.StateDef
|
||||||
|
import hep.dataforge.states.StateDefs
|
||||||
|
import hep.dataforge.states.valueState
|
||||||
|
import hep.dataforge.utils.DateTimeUtils
|
||||||
|
import hep.dataforge.values.ValueType.*
|
||||||
import inr.numass.control.DeviceView
|
import inr.numass.control.DeviceView
|
||||||
import inr.numass.control.magnet.fx.MagnetDisplay
|
import inr.numass.control.magnet.fx.MagnetDisplay
|
||||||
|
import kotlinx.coroutines.runBlocking
|
||||||
import java.time.Duration
|
import java.time.Duration
|
||||||
import java.time.Instant
|
import java.time.Instant
|
||||||
import java.time.temporal.ChronoUnit
|
import java.time.temporal.ChronoUnit
|
||||||
|
@ -16,6 +16,8 @@
|
|||||||
package inr.numass.control.magnet.fx
|
package inr.numass.control.magnet.fx
|
||||||
|
|
||||||
import hep.dataforge.exceptions.PortException
|
import hep.dataforge.exceptions.PortException
|
||||||
|
import hep.dataforge.fx.asDoubleProperty
|
||||||
|
import hep.dataforge.states.ValueState
|
||||||
import inr.numass.control.DeviceDisplayFX
|
import inr.numass.control.DeviceDisplayFX
|
||||||
import inr.numass.control.magnet.LambdaMagnet
|
import inr.numass.control.magnet.LambdaMagnet
|
||||||
import javafx.application.Platform
|
import javafx.application.Platform
|
||||||
@ -24,6 +26,7 @@ import javafx.beans.value.ObservableValue
|
|||||||
import javafx.scene.control.*
|
import javafx.scene.control.*
|
||||||
import javafx.scene.layout.AnchorPane
|
import javafx.scene.layout.AnchorPane
|
||||||
import javafx.scene.paint.Color
|
import javafx.scene.paint.Color
|
||||||
|
import tornadofx.*
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* FXML Controller class
|
* FXML Controller class
|
||||||
|
@ -33,7 +33,7 @@ import hep.dataforge.meta.Meta
|
|||||||
import hep.dataforge.states.StateDef
|
import hep.dataforge.states.StateDef
|
||||||
import hep.dataforge.states.StateDefs
|
import hep.dataforge.states.StateDefs
|
||||||
import hep.dataforge.states.valueState
|
import hep.dataforge.states.valueState
|
||||||
import hep.dataforge.storage.commons.StorageConnection
|
import hep.dataforge.storage.StorageConnection
|
||||||
import hep.dataforge.tables.TableFormatBuilder
|
import hep.dataforge.tables.TableFormatBuilder
|
||||||
import hep.dataforge.tables.ValuesListener
|
import hep.dataforge.tables.ValuesListener
|
||||||
import hep.dataforge.useMeta
|
import hep.dataforge.useMeta
|
||||||
|
@ -4,11 +4,10 @@ import hep.dataforge.control.connections.DeviceConnection
|
|||||||
import hep.dataforge.control.connections.Roles
|
import hep.dataforge.control.connections.Roles
|
||||||
import hep.dataforge.control.devices.Device
|
import hep.dataforge.control.devices.Device
|
||||||
import hep.dataforge.nullable
|
import hep.dataforge.nullable
|
||||||
import hep.dataforge.storage.MutableStorage
|
|
||||||
import hep.dataforge.storage.MutableTableLoader
|
|
||||||
import hep.dataforge.storage.Storage
|
import hep.dataforge.storage.Storage
|
||||||
import hep.dataforge.storage.StorageConnection
|
import hep.dataforge.storage.StorageConnection
|
||||||
import hep.dataforge.storage.files.createTable
|
import hep.dataforge.storage.tables.MutableTableLoader
|
||||||
|
import hep.dataforge.storage.tables.createTable
|
||||||
import hep.dataforge.tables.TableFormat
|
import hep.dataforge.tables.TableFormat
|
||||||
import hep.dataforge.tables.ValuesListener
|
import hep.dataforge.tables.ValuesListener
|
||||||
import hep.dataforge.utils.DateTimeUtils
|
import hep.dataforge.utils.DateTimeUtils
|
||||||
@ -29,11 +28,12 @@ class NumassStorageConnection(private val loaderName: String? = null, private va
|
|||||||
val loaderName = "${loaderName ?: device.name}_$suffix"
|
val loaderName = "${loaderName ?: device.name}_$suffix"
|
||||||
device.forEachConnection(Roles.STORAGE_ROLE, StorageConnection::class.java) { connection ->
|
device.forEachConnection(Roles.STORAGE_ROLE, StorageConnection::class.java) { connection ->
|
||||||
try {
|
try {
|
||||||
|
//create a loader instance for each connected storage
|
||||||
|
val pl = loaderMap.getOrPut(connection.storage) {
|
||||||
|
connection.storage.createTable(loaderName, format)
|
||||||
|
}
|
||||||
|
|
||||||
connection.context.launch(Dispatchers.IO) {
|
connection.context.launch(Dispatchers.IO) {
|
||||||
//create a loader instance for each connected storage
|
|
||||||
val pl = loaderMap.getOrPut(connection.storage) {
|
|
||||||
(connection.storage as MutableStorage).createTable(loaderName, format)
|
|
||||||
}
|
|
||||||
pl.append(point)
|
pl.append(point)
|
||||||
}
|
}
|
||||||
} catch (ex: Exception) {
|
} catch (ex: Exception) {
|
||||||
|
@ -3,7 +3,7 @@ package inr.numass.control
|
|||||||
import hep.dataforge.control.devices.AbstractDevice
|
import hep.dataforge.control.devices.AbstractDevice
|
||||||
import hep.dataforge.nullable
|
import hep.dataforge.nullable
|
||||||
import hep.dataforge.storage.StorageConnection
|
import hep.dataforge.storage.StorageConnection
|
||||||
import hep.dataforge.storage.TableLoader
|
import hep.dataforge.storage.tables.TableLoader
|
||||||
|
|
||||||
import hep.dataforge.values.Values
|
import hep.dataforge.values.Values
|
||||||
import kotlinx.coroutines.Dispatchers
|
import kotlinx.coroutines.Dispatchers
|
||||||
|
@ -2,6 +2,7 @@ package inr.numass.control.readvac
|
|||||||
|
|
||||||
import hep.dataforge.control.devices.Sensor
|
import hep.dataforge.control.devices.Sensor
|
||||||
import hep.dataforge.control.devices.Sensor.Companion.RESULT_VALUE
|
import hep.dataforge.control.devices.Sensor.Companion.RESULT_VALUE
|
||||||
|
import kotlinx.coroutines.runBlocking
|
||||||
import org.apache.commons.cli.DefaultParser
|
import org.apache.commons.cli.DefaultParser
|
||||||
import org.apache.commons.cli.HelpFormatter
|
import org.apache.commons.cli.HelpFormatter
|
||||||
import org.apache.commons.cli.Options
|
import org.apache.commons.cli.Options
|
||||||
|
@ -5,6 +5,10 @@
|
|||||||
*/
|
*/
|
||||||
package inr.numass.control.readvac
|
package inr.numass.control.readvac
|
||||||
|
|
||||||
|
import hep.dataforge.connections.Connection
|
||||||
|
import hep.dataforge.connections.RoleDef
|
||||||
|
import hep.dataforge.context.Context
|
||||||
|
import hep.dataforge.context.launch
|
||||||
import hep.dataforge.control.collectors.RegularPointCollector
|
import hep.dataforge.control.collectors.RegularPointCollector
|
||||||
import hep.dataforge.control.connections.Roles
|
import hep.dataforge.control.connections.Roles
|
||||||
import hep.dataforge.control.devices.Device
|
import hep.dataforge.control.devices.Device
|
||||||
@ -12,9 +16,23 @@ import hep.dataforge.control.devices.DeviceHub
|
|||||||
import hep.dataforge.control.devices.DeviceListener
|
import hep.dataforge.control.devices.DeviceListener
|
||||||
import hep.dataforge.control.devices.PortSensor.Companion.CONNECTED_STATE
|
import hep.dataforge.control.devices.PortSensor.Companion.CONNECTED_STATE
|
||||||
import hep.dataforge.control.devices.Sensor
|
import hep.dataforge.control.devices.Sensor
|
||||||
|
import hep.dataforge.description.ValueDef
|
||||||
import hep.dataforge.exceptions.ControlException
|
import hep.dataforge.exceptions.ControlException
|
||||||
|
import hep.dataforge.meta.Meta
|
||||||
|
import hep.dataforge.names.Name
|
||||||
|
import hep.dataforge.states.StateDef
|
||||||
|
import hep.dataforge.storage.StorageConnection
|
||||||
|
import hep.dataforge.storage.tables.TableLoader
|
||||||
|
import hep.dataforge.storage.tables.createTable
|
||||||
|
import hep.dataforge.tables.TableFormatBuilder
|
||||||
|
import hep.dataforge.utils.DateTimeUtils
|
||||||
|
import hep.dataforge.values.Value
|
||||||
|
import hep.dataforge.values.ValueMap
|
||||||
|
import hep.dataforge.values.ValueType
|
||||||
|
import hep.dataforge.values.Values
|
||||||
import inr.numass.control.DeviceView
|
import inr.numass.control.DeviceView
|
||||||
import inr.numass.control.StorageHelper
|
import inr.numass.control.StorageHelper
|
||||||
|
import kotlinx.coroutines.time.delay
|
||||||
import java.time.Duration
|
import java.time.Duration
|
||||||
import java.time.Instant
|
import java.time.Instant
|
||||||
import java.util.*
|
import java.util.*
|
||||||
@ -77,7 +95,8 @@ class VacCollectorDevice(context: Context, meta: Meta, val sensors: Collection<S
|
|||||||
|
|
||||||
val suffix = DateTimeUtils.fileSuffix()
|
val suffix = DateTimeUtils.fileSuffix()
|
||||||
|
|
||||||
return LoaderFactory.buildPointLoader(connection.storage, "vactms_$suffix", "", "timestamp", format.build())
|
return connection.storage.createTable("vactms_$suffix", format.build())
|
||||||
|
//LoaderFactory.buildPointLoader(connection.storage, "vactms_$suffix", "", "timestamp", format.build())
|
||||||
}
|
}
|
||||||
|
|
||||||
override fun connectAll(connection: Connection, vararg roles: String) {
|
override fun connectAll(connection: Connection, vararg roles: String) {
|
||||||
@ -92,7 +111,7 @@ class VacCollectorDevice(context: Context, meta: Meta, val sensors: Collection<S
|
|||||||
|
|
||||||
|
|
||||||
private fun notifyResult(values: Values, timestamp: Instant = Instant.now()) {
|
private fun notifyResult(values: Values, timestamp: Instant = Instant.now()) {
|
||||||
super.notifyResult(values,timestamp)
|
super.notifyResult(values, timestamp)
|
||||||
helper.push(values)
|
helper.push(values)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -112,7 +131,7 @@ class VacCollectorDevice(context: Context, meta: Meta, val sensors: Collection<S
|
|||||||
while (true) {
|
while (true) {
|
||||||
notifyMeasurementState(MeasurementState.IN_PROGRESS)
|
notifyMeasurementState(MeasurementState.IN_PROGRESS)
|
||||||
sensors.forEach { sensor ->
|
sensors.forEach { sensor ->
|
||||||
if (sensor.states.getBoolean(CONNECTED_STATE,false)) {
|
if (sensor.states.getBoolean(CONNECTED_STATE, false)) {
|
||||||
sensor.measure()
|
sensor.measure()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,50 +1,12 @@
|
|||||||
buildscript {
|
|
||||||
repositories {
|
|
||||||
mavenCentral()
|
|
||||||
}
|
|
||||||
dependencies {
|
|
||||||
classpath 'com.google.protobuf:protobuf-gradle-plugin:0.8.3'
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
apply plugin: 'com.google.protobuf'
|
|
||||||
|
|
||||||
description = "A bse package with minimal dependencies for numass"
|
description = "A bse package with minimal dependencies for numass"
|
||||||
|
|
||||||
|
|
||||||
dependencies {
|
dependencies {
|
||||||
|
compile project(":numass-core:numass-data-api")
|
||||||
|
compile project(":numass-core:numass-data-proto")
|
||||||
compile "hep.dataforge:dataforge-storage"
|
compile "hep.dataforge:dataforge-storage"
|
||||||
compile "hep.dataforge:dataforge-json"
|
compile "hep.dataforge:dataforge-json"
|
||||||
compile 'com.google.protobuf:protobuf-java:3.5.0'
|
|
||||||
|
|
||||||
// https://mvnrepository.com/artifact/com.github.robtimus/sftp-fs
|
// https://mvnrepository.com/artifact/com.github.robtimus/sftp-fs
|
||||||
compile group: 'com.github.robtimus', name: 'sftp-fs', version: '1.1.3'
|
compile group: 'com.github.robtimus', name: 'sftp-fs', version: '1.1.3'
|
||||||
}
|
|
||||||
|
|
||||||
protobuf {
|
|
||||||
// Configure the protoc executable
|
|
||||||
protoc {
|
|
||||||
// Download from repositories
|
|
||||||
artifact = 'com.google.protobuf:protoc:3.5.0'
|
|
||||||
}
|
|
||||||
generatedFilesBaseDir = "$projectDir/gen"
|
|
||||||
}
|
|
||||||
|
|
||||||
compileKotlin {
|
|
||||||
dependsOn(':numass-core:generateProto')
|
|
||||||
}
|
|
||||||
|
|
||||||
sourceSets {
|
|
||||||
main.kotlin.srcDirs += 'gen/main/java'
|
|
||||||
}
|
|
||||||
|
|
||||||
clean {
|
|
||||||
delete protobuf.generatedFilesBaseDir
|
|
||||||
}
|
|
||||||
|
|
||||||
idea {
|
|
||||||
module {
|
|
||||||
sourceDirs += file("${protobuf.generatedFilesBaseDir}/main/java");
|
|
||||||
sourceDirs += file("src/main/proto")
|
|
||||||
}
|
|
||||||
}
|
}
|
20
numass-core/numass-data-api/build.gradle.kts
Normal file
20
numass-core/numass-data-api/build.gradle.kts
Normal file
@ -0,0 +1,20 @@
|
|||||||
|
import org.jetbrains.kotlin.gradle.tasks.KotlinCompile
|
||||||
|
|
||||||
|
plugins {
|
||||||
|
idea
|
||||||
|
kotlin("jvm")
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
repositories {
|
||||||
|
mavenCentral()
|
||||||
|
}
|
||||||
|
|
||||||
|
dependencies {
|
||||||
|
compile(kotlin("stdlib-jdk8"))
|
||||||
|
compile("hep.dataforge:dataforge-core")
|
||||||
|
}
|
||||||
|
|
||||||
|
tasks.withType<KotlinCompile> {
|
||||||
|
kotlinOptions.jvmTarget = "1.8"
|
||||||
|
}
|
@ -1,6 +1,21 @@
|
|||||||
|
/*
|
||||||
|
* Copyright 2018 Alexander Nozik.
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
package inr.numass.data.api
|
package inr.numass.data.api
|
||||||
|
|
||||||
import inr.numass.data.channel
|
|
||||||
import java.io.Serializable
|
import java.io.Serializable
|
||||||
import java.time.Duration
|
import java.time.Duration
|
||||||
import java.time.Instant
|
import java.time.Instant
|
||||||
@ -62,6 +77,8 @@ interface NumassBlock {
|
|||||||
* Stream of frames. Could be empty
|
* Stream of frames. Could be empty
|
||||||
*/
|
*/
|
||||||
val frames: Stream<NumassFrame>
|
val frames: Stream<NumassFrame>
|
||||||
|
|
||||||
|
val channel: Int get() = 0
|
||||||
}
|
}
|
||||||
|
|
||||||
fun OrphanNumassEvent.adopt(parent: NumassBlock): NumassEvent {
|
fun OrphanNumassEvent.adopt(parent: NumassBlock): NumassEvent {
|
@ -1,12 +1,24 @@
|
|||||||
|
/*
|
||||||
|
* Copyright 2018 Alexander Nozik.
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
package inr.numass.data.api
|
package inr.numass.data.api
|
||||||
|
|
||||||
import hep.dataforge.io.envelopes.Envelope
|
|
||||||
import hep.dataforge.meta.Metoid
|
import hep.dataforge.meta.Metoid
|
||||||
import hep.dataforge.providers.Provider
|
import hep.dataforge.providers.Provider
|
||||||
import hep.dataforge.providers.Provides
|
import hep.dataforge.providers.Provides
|
||||||
import inr.numass.data.channel
|
|
||||||
import inr.numass.data.storage.ClassicNumassPoint
|
|
||||||
import inr.numass.data.storage.ProtoNumassPoint
|
|
||||||
import java.time.Duration
|
import java.time.Duration
|
||||||
import java.time.Instant
|
import java.time.Instant
|
||||||
import java.util.stream.Stream
|
import java.util.stream.Stream
|
||||||
@ -119,13 +131,5 @@ interface NumassPoint : Metoid, ParentBlock, Provider {
|
|||||||
const val LENGTH_KEY = "length"
|
const val LENGTH_KEY = "length"
|
||||||
const val HV_KEY = "voltage"
|
const val HV_KEY = "voltage"
|
||||||
const val INDEX_KEY = "index"
|
const val INDEX_KEY = "index"
|
||||||
|
|
||||||
fun read(envelope: Envelope): NumassPoint {
|
|
||||||
return if (envelope.dataType?.startsWith("numass.point.classic") ?: envelope.meta.hasValue("split")) {
|
|
||||||
ClassicNumassPoint(envelope)
|
|
||||||
} else {
|
|
||||||
ProtoNumassPoint.fromEnvelope(envelope)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
48
numass-core/numass-data-proto/build.gradle.kts
Normal file
48
numass-core/numass-data-proto/build.gradle.kts
Normal file
@ -0,0 +1,48 @@
|
|||||||
|
import com.google.protobuf.gradle.GenerateProtoTask
|
||||||
|
import com.google.protobuf.gradle.protobuf
|
||||||
|
import com.google.protobuf.gradle.protoc
|
||||||
|
import org.jetbrains.kotlin.gradle.tasks.KotlinCompile
|
||||||
|
|
||||||
|
plugins {
|
||||||
|
idea
|
||||||
|
kotlin("jvm")
|
||||||
|
id("com.google.protobuf") version "0.8.7"
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
repositories {
|
||||||
|
mavenCentral()
|
||||||
|
}
|
||||||
|
|
||||||
|
dependencies {
|
||||||
|
compile(kotlin("stdlib-jdk8"))
|
||||||
|
compile("com.google.protobuf:protobuf-java:3.6.1")
|
||||||
|
compile(project(":numass-core:numass-data-api"))
|
||||||
|
compile("hep.dataforge:dataforge-storage")
|
||||||
|
}
|
||||||
|
|
||||||
|
tasks.withType<KotlinCompile> {
|
||||||
|
kotlinOptions.jvmTarget = "1.8"
|
||||||
|
dependsOn(":numass-core:numass-data-proto:generateProto")
|
||||||
|
}
|
||||||
|
|
||||||
|
sourceSets{
|
||||||
|
create("proto"){
|
||||||
|
proto {
|
||||||
|
srcDir("src/main/proto")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
protobuf {
|
||||||
|
// Configure the protoc executable
|
||||||
|
protoc {
|
||||||
|
// Download from repositories
|
||||||
|
artifact = "com.google.protobuf:protoc:3.6.1"
|
||||||
|
}
|
||||||
|
generatedFilesBaseDir = "$projectDir/gen"
|
||||||
|
}
|
||||||
|
|
||||||
|
//tasks.getByName("clean").doLast{
|
||||||
|
// delete(protobuf.protobuf.generatedFilesBaseDir)
|
||||||
|
//}
|
@ -1,4 +1,20 @@
|
|||||||
package inr.numass
|
/*
|
||||||
|
* Copyright 2018 Alexander Nozik.
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package inr.numass.data
|
||||||
|
|
||||||
import hep.dataforge.io.envelopes.*
|
import hep.dataforge.io.envelopes.*
|
||||||
import hep.dataforge.values.Value
|
import hep.dataforge.values.Value
|
||||||
@ -105,7 +121,7 @@ class NumassEnvelopeType : EnvelopeType {
|
|||||||
val buffer = it.map(FileChannel.MapMode.READ_ONLY, 0, 6)
|
val buffer = it.map(FileChannel.MapMode.READ_ONLY, 0, 6)
|
||||||
when {
|
when {
|
||||||
//TODO use templates from appropriate types
|
//TODO use templates from appropriate types
|
||||||
buffer.get(0) == '#'.toByte() && buffer.get(1) == '!'.toByte() -> NumassEnvelopeType.INSTANCE
|
buffer.get(0) == '#'.toByte() && buffer.get(1) == '!'.toByte() -> INSTANCE
|
||||||
buffer.get(0) == '#'.toByte() && buffer.get(1) == '!'.toByte() &&
|
buffer.get(0) == '#'.toByte() && buffer.get(1) == '!'.toByte() &&
|
||||||
buffer.get(4) == 'T'.toByte() && buffer.get(5) == 'L'.toByte() -> TaglessEnvelopeType.INSTANCE
|
buffer.get(4) == 'T'.toByte() && buffer.get(5) == 'L'.toByte() -> TaglessEnvelopeType.INSTANCE
|
||||||
buffer.get(0) == '#'.toByte() && buffer.get(1) == '~'.toByte() -> DefaultEnvelopeType.INSTANCE
|
buffer.get(0) == '#'.toByte() && buffer.get(1) == '~'.toByte() -> DefaultEnvelopeType.INSTANCE
|
@ -0,0 +1,52 @@
|
|||||||
|
/*
|
||||||
|
* Copyright 2018 Alexander Nozik.
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package inr.numass.data
|
||||||
|
|
||||||
|
import hep.dataforge.meta.Meta
|
||||||
|
import hep.dataforge.storage.files.MutableFileEnvelope
|
||||||
|
import java.nio.ByteBuffer
|
||||||
|
import java.nio.file.Files
|
||||||
|
import java.nio.file.Path
|
||||||
|
import java.nio.file.StandardOpenOption
|
||||||
|
|
||||||
|
class NumassFileEnvelope(path: Path) : MutableFileEnvelope(path) {
|
||||||
|
|
||||||
|
private val tag by lazy { Files.newByteChannel(path, StandardOpenOption.READ).use { NumassEnvelopeType.LegacyTag().read(it) } }
|
||||||
|
|
||||||
|
override val dataOffset: Long by lazy { (tag.length + tag.metaSize).toLong() }
|
||||||
|
|
||||||
|
override var dataLength: Int
|
||||||
|
get() = tag.dataSize
|
||||||
|
set(value) {
|
||||||
|
if (value > Int.MAX_VALUE) {
|
||||||
|
throw RuntimeException("Too large data block")
|
||||||
|
}
|
||||||
|
tag.dataSize = value
|
||||||
|
if (channel.write(tag.toBytes(), 0L) < tag.length) {
|
||||||
|
throw error("Tag is not overwritten.")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
override val meta: Meta by lazy {
|
||||||
|
val buffer = ByteBuffer.allocate(tag.metaSize).also {
|
||||||
|
channel.read(it, tag.length.toLong())
|
||||||
|
}
|
||||||
|
tag.metaType.reader.readBuffer(buffer)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
@ -1,19 +1,34 @@
|
|||||||
package inr.numass.data.storage
|
/*
|
||||||
|
* Copyright 2018 Alexander Nozik.
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package inr.numass.data
|
||||||
|
|
||||||
import hep.dataforge.context.Context
|
|
||||||
import hep.dataforge.context.Global
|
|
||||||
import hep.dataforge.io.envelopes.Envelope
|
import hep.dataforge.io.envelopes.Envelope
|
||||||
import hep.dataforge.meta.Meta
|
import hep.dataforge.meta.Meta
|
||||||
import inr.numass.data.NumassProto
|
|
||||||
import inr.numass.data.api.*
|
import inr.numass.data.api.*
|
||||||
import inr.numass.data.dataStream
|
|
||||||
import inr.numass.data.legacy.NumassFileEnvelope
|
|
||||||
import org.slf4j.LoggerFactory
|
import org.slf4j.LoggerFactory
|
||||||
|
import java.io.ByteArrayInputStream
|
||||||
|
import java.io.ByteArrayOutputStream
|
||||||
|
import java.io.InputStream
|
||||||
import java.nio.file.Path
|
import java.nio.file.Path
|
||||||
import java.time.Duration
|
import java.time.Duration
|
||||||
import java.time.Instant
|
import java.time.Instant
|
||||||
import java.util.stream.IntStream
|
import java.util.stream.IntStream
|
||||||
import java.util.stream.Stream
|
import java.util.stream.Stream
|
||||||
|
import java.util.zip.Inflater
|
||||||
import kotlin.streams.toList
|
import kotlin.streams.toList
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -61,17 +76,48 @@ class ProtoNumassPoint(override val meta: Meta, val protoBuilder: () -> NumassPr
|
|||||||
return fromEnvelope(NumassFileEnvelope(path))
|
return fromEnvelope(NumassFileEnvelope(path))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get valid data stream utilizing compression if it is present
|
||||||
|
*/
|
||||||
|
private fun Envelope.dataStream(): InputStream = if (this.meta.getString("compression", "none") == "zlib") {
|
||||||
|
//TODO move to new type of data
|
||||||
|
val inflatter = Inflater()
|
||||||
|
val array: ByteArray = with(data.buffer) {
|
||||||
|
if (hasArray()) {
|
||||||
|
array()
|
||||||
|
} else {
|
||||||
|
ByteArray(this.limit()).also {
|
||||||
|
this.position(0)
|
||||||
|
get(it)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
inflatter.setInput(array)
|
||||||
|
val bos = ByteArrayOutputStream()
|
||||||
|
val buffer = ByteArray(8192)
|
||||||
|
while (!inflatter.finished()) {
|
||||||
|
val size = inflatter.inflate(buffer)
|
||||||
|
bos.write(buffer, 0, size)
|
||||||
|
}
|
||||||
|
val unzippeddata = bos.toByteArray()
|
||||||
|
inflatter.end()
|
||||||
|
ByteArrayInputStream(unzippeddata)
|
||||||
|
} else {
|
||||||
|
this.data.stream
|
||||||
|
}
|
||||||
|
|
||||||
fun fromEnvelope(envelope: Envelope): ProtoNumassPoint {
|
fun fromEnvelope(envelope: Envelope): ProtoNumassPoint {
|
||||||
return ProtoNumassPoint(envelope.meta) {
|
return ProtoNumassPoint(envelope.meta) {
|
||||||
envelope.dataStream.use {
|
envelope.dataStream().use {
|
||||||
NumassProto.Point.parseFrom(it)
|
NumassProto.Point.parseFrom(it)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fun readFile(path: String, context: Context = Global): ProtoNumassPoint {
|
// fun readFile(path: String, context: Context = Global): ProtoNumassPoint {
|
||||||
return readFile(context.getFile(path).absolutePath)
|
// return readFile(context.getFile(path).absolutePath)
|
||||||
}
|
// }
|
||||||
|
|
||||||
fun ofEpochNanos(nanos: Long): Instant {
|
fun ofEpochNanos(nanos: Long): Instant {
|
||||||
val seconds = Math.floorDiv(nanos, 1e9.toInt().toLong())
|
val seconds = Math.floorDiv(nanos, 1e9.toInt().toLong())
|
||||||
@ -81,7 +127,7 @@ class ProtoNumassPoint(override val meta: Meta, val protoBuilder: () -> NumassPr
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
class ProtoBlock(val channel: Int, private val block: NumassProto.Point.Channel.Block, val parent: NumassPoint? = null) : NumassBlock {
|
class ProtoBlock(override val channel: Int, private val block: NumassProto.Point.Channel.Block, val parent: NumassPoint? = null) : NumassBlock {
|
||||||
|
|
||||||
override val startTime: Instant
|
override val startTime: Instant
|
||||||
get() = ProtoNumassPoint.ofEpochNanos(block.time)
|
get() = ProtoNumassPoint.ofEpochNanos(block.time)
|
@ -0,0 +1,33 @@
|
|||||||
|
syntax = "proto3";
|
||||||
|
|
||||||
|
package inr.numass.data;
|
||||||
|
|
||||||
|
message Point {
|
||||||
|
// A single channel for multichannel detector readout
|
||||||
|
message Channel {
|
||||||
|
//A continuous measurement block
|
||||||
|
message Block {
|
||||||
|
// Raw data frame
|
||||||
|
message Frame {
|
||||||
|
uint64 time = 1; // Time in nanos from the beginning of the block
|
||||||
|
bytes data = 2; // Frame data as an array of int16 mesured in arbitrary channels
|
||||||
|
}
|
||||||
|
// Event block obtained directly from device of from frame analysis
|
||||||
|
// In order to save space, times and amplitudes are in separate arrays.
|
||||||
|
// Amplitude and time with the same index correspond to the same event
|
||||||
|
message Events {
|
||||||
|
repeated uint64 times = 1; // Array of time in nanos from the beginning of the block
|
||||||
|
repeated uint64 amplitudes = 2; // Array of amplitudes of events in channels
|
||||||
|
}
|
||||||
|
|
||||||
|
uint64 time = 1; // Block start in epoch nanos
|
||||||
|
repeated Frame frames = 2; // Frames array
|
||||||
|
Events events = 3; // Events array
|
||||||
|
uint64 length = 4; // block size in nanos. If missing, take from meta.
|
||||||
|
uint64 bin_size = 5; // tick size in nanos. Obsolete, to be removed
|
||||||
|
}
|
||||||
|
uint64 id = 1; // The number of measuring channel
|
||||||
|
repeated Block blocks = 2; // Blocks
|
||||||
|
}
|
||||||
|
repeated Channel channels = 1; // Array of measuring channels
|
||||||
|
}
|
@ -1,15 +1,27 @@
|
|||||||
|
/*
|
||||||
|
* Copyright 2018 Alexander Nozik.
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
package inr.numass.data
|
package inr.numass.data
|
||||||
|
|
||||||
import hep.dataforge.io.envelopes.Envelope
|
import hep.dataforge.io.envelopes.Envelope
|
||||||
import hep.dataforge.meta.Meta
|
import hep.dataforge.meta.Meta
|
||||||
import hep.dataforge.meta.MetaBuilder
|
import hep.dataforge.meta.MetaBuilder
|
||||||
import inr.numass.data.api.*
|
import inr.numass.data.api.*
|
||||||
import inr.numass.data.storage.ProtoBlock
|
import inr.numass.data.storage.ClassicNumassPoint
|
||||||
import java.io.ByteArrayInputStream
|
|
||||||
import java.io.ByteArrayOutputStream
|
|
||||||
import java.io.InputStream
|
|
||||||
import java.util.stream.Collectors
|
import java.util.stream.Collectors
|
||||||
import java.util.zip.Inflater
|
|
||||||
import kotlin.streams.asSequence
|
import kotlin.streams.asSequence
|
||||||
import kotlin.streams.toList
|
import kotlin.streams.toList
|
||||||
|
|
||||||
@ -42,47 +54,16 @@ object NumassDataUtils {
|
|||||||
fun adapter(): SpectrumAdapter {
|
fun adapter(): SpectrumAdapter {
|
||||||
return SpectrumAdapter("Uset", "CR", "CRerr", "Time")
|
return SpectrumAdapter("Uset", "CR", "CRerr", "Time")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fun read(envelope: Envelope): NumassPoint {
|
||||||
|
return if (envelope.dataType?.startsWith("numass.point.classic") ?: envelope.meta.hasValue("split")) {
|
||||||
|
ClassicNumassPoint(envelope)
|
||||||
|
} else {
|
||||||
|
ProtoNumassPoint.fromEnvelope(envelope)
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Get valid data stream utilizing compression if it is present
|
|
||||||
*/
|
|
||||||
val Envelope.dataStream: InputStream
|
|
||||||
get() = if (this.meta.getString("compression", "none") == "zlib") {
|
|
||||||
//TODO move to new type of data
|
|
||||||
val inflatter = Inflater()
|
|
||||||
val array: ByteArray = with(data.buffer) {
|
|
||||||
if (hasArray()) {
|
|
||||||
array()
|
|
||||||
} else {
|
|
||||||
ByteArray(this.limit()).also {
|
|
||||||
this.position(0)
|
|
||||||
get(it)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
inflatter.setInput(array)
|
|
||||||
val bos = ByteArrayOutputStream()
|
|
||||||
val buffer = ByteArray(8192)
|
|
||||||
while (!inflatter.finished()) {
|
|
||||||
val size = inflatter.inflate(buffer)
|
|
||||||
bos.write(buffer, 0, size)
|
|
||||||
}
|
|
||||||
val unzippeddata = bos.toByteArray()
|
|
||||||
inflatter.end()
|
|
||||||
ByteArrayInputStream(unzippeddata)
|
|
||||||
} else {
|
|
||||||
this.data.stream
|
|
||||||
}
|
|
||||||
|
|
||||||
val NumassBlock.channel: Int
|
|
||||||
get() = if (this is ProtoBlock) {
|
|
||||||
this.channel
|
|
||||||
} else {
|
|
||||||
0
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
suspend fun NumassBlock.transformChain(transform: (NumassEvent, NumassEvent) -> Pair<Short, Long>?): NumassBlock {
|
suspend fun NumassBlock.transformChain(transform: (NumassEvent, NumassEvent) -> Pair<Short, Long>?): NumassBlock {
|
||||||
return SimpleBlock.produce(this.startTime, this.length) {
|
return SimpleBlock.produce(this.startTime, this.length) {
|
||||||
this.events.asSequence()
|
this.events.asSequence()
|
||||||
|
@ -1,70 +0,0 @@
|
|||||||
package inr.numass.data.legacy
|
|
||||||
|
|
||||||
import hep.dataforge.meta.Meta
|
|
||||||
import hep.dataforge.storage.files.MutableFileEnvelope
|
|
||||||
import inr.numass.NumassEnvelopeType
|
|
||||||
import java.nio.ByteBuffer
|
|
||||||
import java.nio.file.Files
|
|
||||||
import java.nio.file.Path
|
|
||||||
import java.nio.file.StandardOpenOption
|
|
||||||
|
|
||||||
class NumassFileEnvelope(path: Path) : MutableFileEnvelope(path) {
|
|
||||||
|
|
||||||
private val tag by lazy { Files.newByteChannel(path, StandardOpenOption.READ).use { NumassEnvelopeType.LegacyTag().read(it) } }
|
|
||||||
|
|
||||||
override val dataOffset: Long by lazy { (tag.length + tag.metaSize).toLong() }
|
|
||||||
|
|
||||||
override var dataLength: Int
|
|
||||||
get() = tag.dataSize
|
|
||||||
set(value) {
|
|
||||||
if (value > Int.MAX_VALUE) {
|
|
||||||
throw RuntimeException("Too large data block")
|
|
||||||
}
|
|
||||||
tag.dataSize = value
|
|
||||||
if (channel.write(tag.toBytes(), 0L) < tag.length) {
|
|
||||||
throw error("Tag is not overwritten.")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
override val meta: Meta by lazy {
|
|
||||||
val buffer = ByteBuffer.allocate(tag.metaSize).also {
|
|
||||||
channel.read(it, tag.length.toLong())
|
|
||||||
}
|
|
||||||
tag.metaType.reader.readBuffer(buffer)
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
//
|
|
||||||
// override fun buildTag(): EnvelopeTag {
|
|
||||||
// return NumassEnvelopeType.LegacyTag()
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// companion object {
|
|
||||||
//
|
|
||||||
// val LEGACY_START_SEQUENCE = byteArrayOf('#'.toByte(), '!'.toByte())
|
|
||||||
// val LEGACY_END_SEQUENCE = byteArrayOf('!'.toByte(), '#'.toByte(), '\r'.toByte(), '\n'.toByte())
|
|
||||||
//
|
|
||||||
// fun open(path: Path, readOnly: Boolean): MutableFileEnvelope {
|
|
||||||
// // if (!Files.exists(path)) {
|
|
||||||
// // throw new RuntimeException("File envelope does not exist");
|
|
||||||
// // }
|
|
||||||
//
|
|
||||||
// try {
|
|
||||||
// FileChannel.open(path, READ).use { channel ->
|
|
||||||
// val buffer = channel.map(FileChannel.MapMode.READ_ONLY, 0, 2)
|
|
||||||
// return if (buffer.compareTo(ByteBuffer.wrap(LEGACY_START_SEQUENCE)) == 0) {
|
|
||||||
// NumassFileEnvelope(path, readOnly)
|
|
||||||
// } else {
|
|
||||||
// MutableFileEnvelope.open(path, readOnly)
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// } catch (e: IOException) {
|
|
||||||
// throw RuntimeException("Failed to open file envelope", e)
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
|
|
@ -2,11 +2,11 @@ package inr.numass.data.storage
|
|||||||
|
|
||||||
import hep.dataforge.io.envelopes.Envelope
|
import hep.dataforge.io.envelopes.Envelope
|
||||||
import hep.dataforge.meta.Meta
|
import hep.dataforge.meta.Meta
|
||||||
|
import inr.numass.data.NumassFileEnvelope
|
||||||
import inr.numass.data.api.NumassBlock
|
import inr.numass.data.api.NumassBlock
|
||||||
import inr.numass.data.api.NumassEvent
|
import inr.numass.data.api.NumassEvent
|
||||||
import inr.numass.data.api.NumassFrame
|
import inr.numass.data.api.NumassFrame
|
||||||
import inr.numass.data.api.NumassPoint
|
import inr.numass.data.api.NumassPoint
|
||||||
import inr.numass.data.legacy.NumassFileEnvelope
|
|
||||||
import org.slf4j.LoggerFactory
|
import org.slf4j.LoggerFactory
|
||||||
import java.io.IOException
|
import java.io.IOException
|
||||||
import java.nio.ByteBuffer
|
import java.nio.ByteBuffer
|
||||||
|
@ -25,7 +25,8 @@ import hep.dataforge.storage.Loader
|
|||||||
import hep.dataforge.storage.StorageElement
|
import hep.dataforge.storage.StorageElement
|
||||||
import hep.dataforge.storage.files.FileStorageElement
|
import hep.dataforge.storage.files.FileStorageElement
|
||||||
import hep.dataforge.tables.Table
|
import hep.dataforge.tables.Table
|
||||||
import inr.numass.NumassEnvelopeType
|
import inr.numass.data.NumassDataUtils
|
||||||
|
import inr.numass.data.NumassEnvelopeType
|
||||||
import inr.numass.data.api.NumassPoint
|
import inr.numass.data.api.NumassPoint
|
||||||
import inr.numass.data.api.NumassSet
|
import inr.numass.data.api.NumassSet
|
||||||
import org.slf4j.LoggerFactory
|
import org.slf4j.LoggerFactory
|
||||||
@ -93,7 +94,7 @@ class NumassDataLoader(
|
|||||||
|
|
||||||
override val points: List<NumassPoint>
|
override val points: List<NumassPoint>
|
||||||
get() = pointEnvelopes.map {
|
get() = pointEnvelopes.map {
|
||||||
NumassPoint.read(it)
|
NumassDataUtils.read(it)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -23,7 +23,7 @@ import hep.dataforge.meta.Meta
|
|||||||
import hep.dataforge.storage.StorageElement
|
import hep.dataforge.storage.StorageElement
|
||||||
import hep.dataforge.storage.files.FileStorage
|
import hep.dataforge.storage.files.FileStorage
|
||||||
import hep.dataforge.storage.files.FileStorageElement
|
import hep.dataforge.storage.files.FileStorageElement
|
||||||
import inr.numass.NumassEnvelopeType
|
import inr.numass.data.NumassEnvelopeType
|
||||||
import kotlinx.coroutines.runBlocking
|
import kotlinx.coroutines.runBlocking
|
||||||
import java.nio.file.Files
|
import java.nio.file.Files
|
||||||
import java.nio.file.Path
|
import java.nio.file.Path
|
||||||
|
@ -1 +1 @@
|
|||||||
inr.numass.NumassEnvelopeType
|
inr.numass.data.NumassEnvelopeType
|
@ -21,10 +21,10 @@ import hep.dataforge.description.Descriptors
|
|||||||
import hep.dataforge.meta.buildMeta
|
import hep.dataforge.meta.buildMeta
|
||||||
import hep.dataforge.plots.data.DataPlot
|
import hep.dataforge.plots.data.DataPlot
|
||||||
import inr.numass.NumassPlugin
|
import inr.numass.NumassPlugin
|
||||||
|
import inr.numass.data.ProtoNumassPoint
|
||||||
import inr.numass.data.analyzers.NumassAnalyzer.Companion.AMPLITUDE_ADAPTER
|
import inr.numass.data.analyzers.NumassAnalyzer.Companion.AMPLITUDE_ADAPTER
|
||||||
import inr.numass.data.analyzers.SmartAnalyzer
|
import inr.numass.data.analyzers.SmartAnalyzer
|
||||||
import inr.numass.data.analyzers.withBinning
|
import inr.numass.data.analyzers.withBinning
|
||||||
import inr.numass.data.storage.ProtoNumassPoint
|
|
||||||
import inr.numass.displayChart
|
import inr.numass.displayChart
|
||||||
import java.nio.file.Paths
|
import java.nio.file.Paths
|
||||||
|
|
||||||
|
@ -12,7 +12,6 @@ import inr.numass.data.analyzers.TimeAnalyzer
|
|||||||
import inr.numass.data.api.NumassPoint
|
import inr.numass.data.api.NumassPoint
|
||||||
import inr.numass.data.api.NumassSet
|
import inr.numass.data.api.NumassSet
|
||||||
import inr.numass.data.api.SimpleNumassPoint
|
import inr.numass.data.api.SimpleNumassPoint
|
||||||
import inr.numass.data.channel
|
|
||||||
import inr.numass.data.storage.NumassDirectory
|
import inr.numass.data.storage.NumassDirectory
|
||||||
|
|
||||||
fun main(args: Array<String>) {
|
fun main(args: Array<String>) {
|
||||||
|
@ -1,8 +1,7 @@
|
|||||||
package inr.numass.scripts.tristan
|
package inr.numass.scripts.tristan
|
||||||
|
|
||||||
import inr.numass.data.channel
|
import inr.numass.data.ProtoNumassPoint
|
||||||
import inr.numass.data.plotAmplitudeSpectrum
|
import inr.numass.data.plotAmplitudeSpectrum
|
||||||
import inr.numass.data.storage.ProtoNumassPoint
|
|
||||||
import inr.numass.data.transformChain
|
import inr.numass.data.transformChain
|
||||||
import kotlinx.coroutines.runBlocking
|
import kotlinx.coroutines.runBlocking
|
||||||
import java.io.File
|
import java.io.File
|
||||||
|
@ -20,7 +20,6 @@ import hep.dataforge.context.Global
|
|||||||
import hep.dataforge.storage.files.FileStorage
|
import hep.dataforge.storage.files.FileStorage
|
||||||
import hep.dataforge.toList
|
import hep.dataforge.toList
|
||||||
import inr.numass.data.api.NumassPoint
|
import inr.numass.data.api.NumassPoint
|
||||||
import inr.numass.data.channel
|
|
||||||
import inr.numass.data.storage.NumassDataLoader
|
import inr.numass.data.storage.NumassDataLoader
|
||||||
import inr.numass.data.storage.NumassDirectory
|
import inr.numass.data.storage.NumassDirectory
|
||||||
|
|
||||||
|
@ -1,11 +1,10 @@
|
|||||||
package inr.numass.scripts.tristan
|
package inr.numass.scripts.tristan
|
||||||
|
|
||||||
|
import inr.numass.data.ProtoNumassPoint
|
||||||
import inr.numass.data.api.MetaBlock
|
import inr.numass.data.api.MetaBlock
|
||||||
import inr.numass.data.api.NumassBlock
|
import inr.numass.data.api.NumassBlock
|
||||||
import inr.numass.data.api.NumassPoint
|
import inr.numass.data.api.NumassPoint
|
||||||
import inr.numass.data.channel
|
|
||||||
import inr.numass.data.plotAmplitudeSpectrum
|
import inr.numass.data.plotAmplitudeSpectrum
|
||||||
import inr.numass.data.storage.ProtoNumassPoint
|
|
||||||
import java.io.File
|
import java.io.File
|
||||||
|
|
||||||
|
|
||||||
|
@ -8,8 +8,8 @@ import hep.dataforge.fx.runGoal
|
|||||||
import hep.dataforge.fx.ui
|
import hep.dataforge.fx.ui
|
||||||
import hep.dataforge.storage.Storage
|
import hep.dataforge.storage.Storage
|
||||||
import inr.numass.NumassProperties
|
import inr.numass.NumassProperties
|
||||||
import inr.numass.data.api.NumassPoint
|
import inr.numass.data.NumassDataUtils
|
||||||
import inr.numass.data.legacy.NumassFileEnvelope
|
import inr.numass.data.NumassFileEnvelope
|
||||||
import inr.numass.data.storage.NumassDataLoader
|
import inr.numass.data.storage.NumassDataLoader
|
||||||
import inr.numass.data.storage.NumassDirectory
|
import inr.numass.data.storage.NumassDirectory
|
||||||
import javafx.beans.property.SimpleObjectProperty
|
import javafx.beans.property.SimpleObjectProperty
|
||||||
@ -183,7 +183,7 @@ class MainView(val context: Context = Global.getContext("viewer")) : View(title
|
|||||||
envelope?.let {
|
envelope?.let {
|
||||||
if (it.meta.hasMeta("external_meta")) {
|
if (it.meta.hasMeta("external_meta")) {
|
||||||
//try to read as point
|
//try to read as point
|
||||||
val point = NumassPoint.read(it)
|
val point = NumassDataUtils.read(it)
|
||||||
runLater {
|
runLater {
|
||||||
contentView = AmplitudeView().apply {
|
contentView = AmplitudeView().apply {
|
||||||
set(path.fileName.toString(), CachedPoint(point))
|
set(path.fileName.toString(), CachedPoint(point))
|
||||||
|
@ -8,7 +8,8 @@ import hep.dataforge.fx.ui
|
|||||||
import hep.dataforge.plots.PlotGroup
|
import hep.dataforge.plots.PlotGroup
|
||||||
import hep.dataforge.plots.data.DataPlot
|
import hep.dataforge.plots.data.DataPlot
|
||||||
import hep.dataforge.plots.jfreechart.JFreeChartFrame
|
import hep.dataforge.plots.jfreechart.JFreeChartFrame
|
||||||
import hep.dataforge.storage.TableLoader
|
import hep.dataforge.storage.tables.TableLoader
|
||||||
|
import hep.dataforge.storage.tables.asTable
|
||||||
import hep.dataforge.tables.Adapters
|
import hep.dataforge.tables.Adapters
|
||||||
import hep.dataforge.tables.Table
|
import hep.dataforge.tables.Table
|
||||||
import javafx.collections.FXCollections
|
import javafx.collections.FXCollections
|
||||||
@ -68,7 +69,7 @@ class SlowControlView : View(title = "Numass slow control view", icon = ImageVie
|
|||||||
|
|
||||||
private suspend fun getData(loader: TableLoader): Table {
|
private suspend fun getData(loader: TableLoader): Table {
|
||||||
//TODO add query
|
//TODO add query
|
||||||
return loader.asTable()
|
return loader.asTable().await()
|
||||||
}
|
}
|
||||||
|
|
||||||
operator fun set(id: String, loader: TableLoader) {
|
operator fun set(id: String, loader: TableLoader) {
|
||||||
|
@ -6,8 +6,8 @@ import hep.dataforge.meta.Meta
|
|||||||
import hep.dataforge.meta.Metoid
|
import hep.dataforge.meta.Metoid
|
||||||
import hep.dataforge.names.AlphanumComparator
|
import hep.dataforge.names.AlphanumComparator
|
||||||
import hep.dataforge.storage.Storage
|
import hep.dataforge.storage.Storage
|
||||||
import hep.dataforge.storage.TableLoader
|
|
||||||
import hep.dataforge.storage.files.FileTableLoader
|
import hep.dataforge.storage.files.FileTableLoader
|
||||||
|
import hep.dataforge.storage.tables.TableLoader
|
||||||
import inr.numass.data.api.NumassPoint
|
import inr.numass.data.api.NumassPoint
|
||||||
import inr.numass.data.api.NumassSet
|
import inr.numass.data.api.NumassSet
|
||||||
import inr.numass.data.storage.NumassDataLoader
|
import inr.numass.data.storage.NumassDataLoader
|
||||||
|
@ -1,20 +1,25 @@
|
|||||||
rootProject.name = 'numass'
|
rootProject.name = 'numass'
|
||||||
|
|
||||||
include ":numass-control"
|
include ":numass-control"
|
||||||
include ":numass-control:cryotemp"
|
include ":numass-control:cryotemp"
|
||||||
include ":numass-control:magnet"
|
include ":numass-control:magnet"
|
||||||
include ":numass-control:msp"
|
include ":numass-control:msp"
|
||||||
include ":numass-control:vac"
|
include ":numass-control:vac"
|
||||||
//include ":numass-control:control-room"
|
//include ":numass-control:control-room"
|
||||||
include ":numass-control:dante"
|
include ":numass-control:dante"
|
||||||
//
|
//
|
||||||
include ":numass-main"
|
include ":numass-main"
|
||||||
//
|
//
|
||||||
include ":numass-core"
|
include ":numass-core"
|
||||||
//include ":numass-server"
|
|
||||||
//
|
include 'numass-core:numass-data-api'
|
||||||
include ":numass-viewer"
|
include 'numass-core:numass-data-proto'
|
||||||
|
|
||||||
if(file("../dataforge").exists()){
|
//include ":numass-server"
|
||||||
includeBuild('../dataforge')
|
//
|
||||||
}
|
include ":numass-viewer"
|
||||||
|
|
||||||
|
if (file("../dataforge").exists()) {
|
||||||
|
includeBuild('../dataforge')
|
||||||
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user