-
Hi all, I'm working on optimizing a multi-objective biological process where experimental evaluations are extremely expensive. To mitigate costs, we initially conducted a space-filling design to obtain some high-fidelity experimental data. Additionally, we have a mathematical model of the process, which, while less reliable than real experiments, is orders of magnitude cheaper to evaluate (about 100,000x cheaper). Goal I want to generate new experimental candidates in parallel (multiple candidates) that maximize the process outcome, but I do not want to suggest low-fidelity evaluations. The model should only be used to inform and improve our high-fidelity experimental selection. Proposed Approach I’m currently considering the following:
Questions:
Many thanks for your insights! |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 9 replies
-
Hi @EvanClaes, a general reference for multi-objective multi-fidelity (esp. within botorch) is our 2023 ICML paper on the Hypervolume Knowledge Gradient: https://proceedings.mlr.press/v202/daulton23a.html. There is also a tutorial on decoupled MOBO (thought that may not be 100% relevant in your context where you can "precompute" a bunch of low fidelity evaluations): https://botorch.org/docs/tutorials/decoupled_mobo/
IIUC this means that there are two discrete fidelities: costly full eval and mathematical model (but no feature that would describe the level of approximation), right?
With that you can follow the rest of your proposed approach:
where the fixed feature just is the high-fidelity task index. Hope this helps - let us know if you have additional questions. |
Beta Was this translation helpful? Give feedback.
-
Hi Evan,
The main thing that the ICM kernel (MuliTaskGP) is that it's going to help
with the shared assumptions re: how important each input is (the kernel
lengthscale). It will also borrow strength across the data and can handle
various linear or potentially non-linear correlations. As a rule of thumb,
you want the pearson correlation between the tasks to be at least 0.4 or so.
https://jmlr.org/papers/volume20/18-225/18-225.pdf provides a fairly
self-contained tutorial and intuition around MTGPs for multi-fidelity
modeling (see e.g., S3.1 and particularly S6).
Best,
e
…On Mon, Mar 3, 2025 at 12:16 PM EvanClaes ***@***.***> wrote:
Hello Max,
Thank you so much for this feedback. I indeed have discrete fidelities, so
your proposed solution makes sense.
Maybe one more question on the low fidelity data. The process model we
have captures the dynamics of the real process relatively well, although
the predictions are sometimes off by a certain factor. We can calibrate the
model with this factor. Would you say that this is important for the
MultiTaskGP (does it care about absolute values), or is it sufficient that
the general trends/dynamics are represented accurately?
Would you mind leaving this open for a couple more days, in case I
encounter some issues with the implementation? I can probably try this out
by the beginning of next week.
Have a great day!
—
Reply to this email directly, view it on GitHub
<#2758 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAAW34PUAYE7TFV4X7RQ5J32SSE7HAVCNFSM6AAAAABYGR3Y5WVHI2DSMVQWIX3LMV43URDJONRXK43TNFXW4Q3PNVWWK3TUHMYTEMZXHA3DINY>
.
You are receiving this because you are subscribed to this thread.Message
ID: ***@***.***>
|
Beta Was this translation helpful? Give feedback.
Hello Max,
I solved the second issue (with all candidates being the same); I had an issue with my constraint.
I can still provide a toy example if you suspect the warning message may be problematic. Otherwise, you can close this topic. To me the candidates that are being produced make sense.
Thanks for all the help!