2025-02-05 graphs lecture 5

Data

2025-02-05

Table of Contents

 

Summary

1. Graph Signals and Graph Signal Processing

Graph filters work ok for many things, but are limited to linear representations. Therefore, they lack expressivity if the task is complicated. ie, since these operations are only linear, we not able to represent more complex relationships.

Thus, we need to add nonlinearities (ex. ReLU etc) to help with the expressivity of our model. These techniques have been implemented in CNNs for image processing and other domains. GNNs will extend the ideas seen in CNNs to the graph domain - graph convolution .

Simple CNN

We can also think of any image as a graph, where the graph signals are the RGB values at each pixel, and the graph structure is determined by the pixel layout/grid of the image.

Graph Perceptron

Let . The graph perceptron is defined as

&\to_{u} y \cdot \sigma(u) \;\;\text{pointwise nonlinearity} \to_{y}

\end{aligned}$$

ie a graph convolution followed by a pointwise nonlinear function

Examples

  • ReLU
  • Sigmoid
  • hypertangent

(see graph perceptron)

Question

Is a graph perceptron local?

Yes, we can still write the perceptron in terms of the local nodes, since

Multi-Layer Graph Perceptron

A(n -layer) multi-layer graph perceptron (MGLP) is simply several graph perceptrons in sequence. This forms a function with input , and output .

Where at each layer, We call the -layer embedding.

For conciseness, we define our weights and represent our MLGP as

Illustration

(see multi-layer graph perceptron)

Full-Fledged GNNs

Real-world graph data is often high dimensional.

  • in addition to having signal values at the nodes, we often have many features. That is, our signal where is the number of nodes and is the number of features
  • (previously we have been looking at the case where )
  • Thus it will be necessary for the embeddings to be multi-dimensional. Note that we also may want embeddings to be multi-dimensional for higher expressivity/flexibility/better representations, not solely because it is necessary.

Example

Suppose are drones communicating via wifi and is their spatial coordinate in 3D: .

In order to do this, we need graph convolution that take in multi-dimensional data.

Convolutional Filter Bank

Let , for . Then a convolutional filter bank (or multiple input multiple output graph convolution) is a multi-dimensional graph convolutiongiven by

  • is still a graph diffusion/shift
  • is a linear transformation mapping features to features

(see convolutional filter bank)

GNN

A “full-fledged” GNN layer is given by for , where

Here, is still called an -layer embedding.

(see Graph Neural Networks)