Multi-Task & Multi-Fidelity modeling #2388
Replies: 2 comments 2 replies
-
Ok while writing this I realized that no cost-aware acquisition function for the described problem is necessary. I can use a FixedFeatureAcquisition function to fix the fidelity_feature column to the value of the highest fidelity and then use any acquisition function. But the question still prevails: should you prefer any model over the other in this setup? |
Beta Was this translation helpful? Give feedback.
-
A multi-fidelity model should work better when the fidelities are ordered, so you know that the lower-fidelity observations are worse than the higher-fidelity observations, not just different. But since you've fit both, you can look at the data and not rely on theory, by checking which has better cross-validation performance. Ideally, you'd do that on the target fidelity, if there is enough data on it. |
Beta Was this translation helpful? Give feedback.
-
Hi i am
Currently trying to do a multi-fidelity optimization. I was trying it with the SingleTaskMultiFidelityGP and now i am trying it with the MultiTaskGP.
The problem setup is actually this:
Observations at different discrete fidelity levels are available but new candidates form the BO should only be inferred from the highest fidelity.
I am using both the SingleTaskMultiFidelityGP and MultiTaskGP to have a "knowledge transfer" between the data fidelities.
For the setup of the BO with the SingleTaskMultiFidelityGP was following the discrete_multi_fidelity_bo tutorial. The whole setup seems kind of over the top with the definition of a cost model and so on. I adjusted it to only inferre new candidates in each BO iteration for the highest fidelity.
Here a quick code snipet form the adjusted code:
Assuming that the
Testfunction
has the correct attributes.For the MulitTaskGP I defined a GenericMCObjective that selects only the output of the highest fidelity.
For that problem definition, what model is the better choice?
To be honest I don't 100% get the DownsamplingKernel and the ExponentialDecayKernel from the original paper of the SingleTaskMultiFidelityGP. What exactly are the iteration fidelity parameters?
Is the SingleTaskMultiFidelityGP despite the rather complex set-up better for this problem setup than the MultiTaskGP because of the kernel structure?
I was also considering doing a polynomial or any other function fit of the low-fidelity data and then to introduce it to a SingleTaskGP as Prior Mean function. Might this be better than the other modeling choices?
I know these are more general questions, especially the last one and it is dependent on the data and dimensionality of the problem and so on but i am thankful for any help or recommendations.
Best regards,
Stefan Tönnis
Beta Was this translation helpful? Give feedback.
All reactions