AIML-K / GNN_Survey

updating papers related to GNN
0 stars 0 forks source link

Bayesian optimisation of functions on graphs #4

Open 2nazero opened 1 month ago

2nazero commented 1 month ago

Bayesian optimisation of functions on graphs

@article{wan2023bayesian,
  title={Bayesian optimisation of functions on graphs},
  author={Wan, Xingchen and Osselin, Pierre and Kenlay, Henry and Ru, Binxin and Osborne, Michael A and Dong, Xiaowen},
  journal={Advances in Neural Information Processing Systems},
  volume={36},
  pages={43012--43040},
  year={2023}
}
2nazero commented 1 month ago

A Very Short Summary

To sum up this paper in one paragraph, this paper is trying to tell "the need for efficient optimization techniques on graph-structured data." Existing methods either ignore function information or are impractical for large-scale graphs. The authors propose a Bayesian optimization framework that overcomes these limitations.

Important Framework

스크린샷 2024-10-17 오후 12 55 25

This figure is the main illustration of what this paper is trying to propose. Here is the explanation!

(a) Start with a Local Subgraph

(b) Decide the Next Node to Query

(c) Update and Move to a New Subgraph

2nazero commented 1 month ago

Preliminaries

BO uses a statistical surrogate model* to approximate the objective function and an acquisition function $$α(x)$$ to balance exploitation and exploration under the principle of optimism in the face of uncertainty.

Surrogate model*

Explanation of Surrogate Model

The optimization task is to find the node(s) $$v^*$$ that minimizes the objective function:

$$ v^* = \arg \min_{v \in V} f(v) $$

(@kyungheee) I'm kind of confused here. So does "minimizing" objective function literally means finding the most smallest function value of the objective function?

Bayesian Optimisation on Graph

There are several challenges of applying BO to Graphs which are:

Algorithm of BayesOptG -1

스크린샷 2024-10-17 오후 2 37 43

Algorithm of BayesOptG -2 (Step 8)

스크린샷 2024-10-17 오후 2 47 02

Tractable Optimisation via Local Modelling

Solutions to the two problems: 1) Large Graphs 2) Incomplete Graph Information

Proposed Solution: 1) Local Subgraph Focus Instead of working with the entire graph (which is computationally expensive), they focus on a small subset of nodes near the best node found so far. 2) Step-by-step Process

2nazero commented 1 month ago

Relation to Trust-region BO Method (The Key Difference)

  1. Custom Distance Metric:

    • Unlike traditional trust region methods, BayesOptG uses a bespoke distance metric designed specifically for the graph space. This custom metric accounts for the unique structure of graphs and how nodes are connected.
  2. Handling Imperfect Graph Knowledge:

    • In traditional trust region BO, the main goal is to avoid over-exploration in high-dimensional spaces. However, in BayesOptG, local subgraphs are crucial for dealing with incomplete knowledge of the graph structure. Instead of needing the entire graph upfront, BayesOptG only needs to reveal the subgraph's topology at each iteration, making it efficient for handling partial information.
  3. Improved Scalability:

    • Using trust regions also improves the scalability of BayesOptG. By focusing on small, local subgraphs instead of the entire graph, computational costs are reduced, resulting in a massive speed-up, as illustrated in Fig. 3. 스크린샷 2024-10-17 오후 3 28 06
2nazero commented 1 month ago

Experiments and the Results

스크린샷 2024-10-17 오후 3 38 44 스크린샷 2024-10-17 오후 3 39 07 스크린샷 2024-10-17 오후 3 41 27 스크린샷 2024-10-17 오후 3 41 39 스크린샷 2024-10-17 오후 3 41 55 스크린샷 2024-10-17 오후 3 42 05
kyungheee commented 1 month ago

@2nazero

It seems like your question is asking whether an optimal minimum always exist. In other words, you seem to be considering the possibility that it may or may not converge to a specific value

Bayesian optimization typically assumes that the objective function is bounded in the region where the optimization is conducted. because the Gaussian process assumes that the value of a function will follow a normal distribution for a given input, however If the function diverges, the Gaussian process would fail to model it properly, rendering the optimization meaningless.

Unbounded Bayesian Optimization via Regularization

Additionally, I found a paper that I haven't read yet but that seems relevant. Let's read together