Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to handle multiple discontinuous data sets. #27

Open
taylorh140 opened this issue May 13, 2021 · 2 comments
Open

How to handle multiple discontinuous data sets. #27

taylorh140 opened this issue May 13, 2021 · 2 comments

Comments

@taylorh140
Copy link

Many experiments may only be performed in short operational bursts. To gain a better idea of the system behavior multiple runs can be used to help gain assurance, and provide better model fitting. Does this library support data, if it does how should the data be formatted?

@CPCLAB-UNIPI
Copy link
Owner

CPCLAB-UNIPI commented May 14, 2021

Dear Taylor,
we are not sure to understand well your question.
SIPPY is designed to take as input, data obtained from different sources (experimental set-up, machines, equipment, plants, or simulation).
Take always the user guide (here in attach) as a reference for the right data format.
If you want to run multiple identification analyses for repeated data sets of the same system, you can embed the call at SIPPY module in a for loop and store the different solutions for post-processing analysis.
Please let us know more in details what is your specific request and we could be more helpful.
Thanks,
The SYPPY team

@jamestjsp
Copy link

jamestjsp commented Aug 1, 2021

I think the question is "How to select data segments form different experiments?". Some methods have discussed in this book.

System Identification Theory for the User: Second Edition, Lennart Ljung, Linkoping, University Sweden

14.3 SELECTING SEGMENTS OF DATA AND MERGING EXPERIMENTS

When data from an identification experiment or in particular, from normal operating records are plotted, it often happens that there are portions of bad data or nonrelevant information. The reason could be that there are long stretches of missing data which will be difficult or computationally costly to reconstruct. There could be portions with disturbances that are considered to be non-representative, or that take the process into operating points that are of less interest. In particular for normal operating records, there could also be long periods of "no information:" nothing seems to happen that carries any information about the/process dynamics. In these cases it is natural to select segments of the original data set which are considered to contain relevant information about dynamics of interest. The procedure of how to select such segments will basically be subjective and will have to rely mostly upon intuition and process insights.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants