Skip to content

Experimental Data Redundancy

Miguel Tomas Silva edited this page Mar 25, 2023 · 26 revisions

Experimental Data Redundancy during an experiment is made on every dataset uploaded to a data repository, for instance, a dataverse, where previously uploaded dataset values are reuploaded on each subsequent dataset with newer data. On the repository side, as an experiment advances, the number of dataset files available increases, with overlapping measured sensor data values between consecutive files.

The last dataset upload holds all data values with a unique fingerprint ID and is linked together in a blockchain-like kind of logic.

It is important to mention here, the importance of linear timed dataset uploads as a way to assist in data acquisition validation of experimental data origins. Therefore is important to happen:

  • TRUE randomized dataset uploads to a data repository, by the smart DAQ setup in any experiment and during the course of an experimental campaign
  • TRUE randomized dataset uploads requests made remotely by the data repository.

During the day, at work hours, and during the night when all is dark.


Collaborative Smart DAQ

Is also possible to set up many Smart DAQs to collaborate together autonomously. For instance, in an experimental setup, with the same batch of specimens to be tested simultaneously, each with its own smart DAQ collecting sensor data, DAQs can be configured to exchange experimental data among each, and do:

  • synchronization of collected data among each other, for instance, using the same Primary key as a common index for the different databases holding experimental data
  • Link individual measurements on each individual specimen, together, by sharing individual fingerprint data IDs with each other stored redundantly but with the same index key across the different local databases.

This blockchain-like logic includes not only the fingerprint ID of the previous data measurement but also all fingerprint IDs generated on all other specimens in the same experimental setup. An example, let's say a researcher has set up 3 specimens for a round of testing, each scheduled sensor data measured, on each of the 3 specimens, will include the sensor data itself with its unique fingerprint ID, and also all fingerprint IDs from the other 2 specimens from the current time-indexed measurement and also from the previous measurement. In total, for this particular example, any sensor data, at any given time, will hold 6 Unique fingerprint IDs on every individual "experimental data block".

Clone this wiki locally