User Tools

Site Tools


public:grl_readingmemo

Reading memo for GRL

* Book: GRL (Graph Representation Learning) by W.L. Hamilton (2020)

General points for thorough reading (not extensive reading)

  • Read every words and mark/memorize important words.
  • Try to understand every piece the authors want to tell us - if needed, read other reference(s).
  • However, if you cannot understand something after you have tried more than 30 minutes, discuss it with others.
  • Try to use low-entropy description in explaining.

#01, liangz

range: from the title page to p8 (end of Chapter 1).

NOTICE: please read both the scanned pictures and this memo.

i: title page

  • official book
  • author William L. Hamilton was an Assistant Professor at that time at McGill Univ.
  • McGill Univ., a famous university in Canada, 12 Nobel laureates, and graduate Yoshua Bengio (2018 ACM Turing Award, one of the three Godfathers of deep learning)

ii: Abstract page

  • no-free-lunch theorem
  • ⇒ inductive bias is important before optimization (and before machine learning). Ex. Consider to use a linear regression to analyze a dataset. The assumption that a linear regression is approporiate is an inductive bias.
  • What is induction?
  • Lin & Tegmark'16 Why does deep and cheap learning work so well? ⇒ Assumptions: (1) Low polynomial order in the real world (2) Locality of data (3) Symmetry phenomenon.
  • 3 graph learning topics: (1) embedding (2) CNN → Graph (3) Message-passing approach

iii-v: Contents

  • not important now

vi: Preface

  • “past seven years” ⇒ graph learning started from 2013
  • at 2020, fastest growing sub-areas of deep learning
  • audience: shall have some background in machine learning and deep learning (e.g., Goodfellow et al, 2016)

vii-viii: Acknowledgements

  • connections of the author (e.g., Jure Leskovec is a famous researcher in network science)

p1: Chapter 1 Introduction

  • node, edge, relations, graph ⇒ ask the audience to illustrate some graphs
  • Zachary Karate Club Network (1977) and on its importance

p2:

  • It mentioned “a dramatic increase in the quantity and quality of graph data in the last 25 years.” ⇒ Why 25 years? (hint: It means since 1995)
  • ML is not the only way but may be interesting.
  • adjacency matrix, adjacency list, simple graph and {0,1}
  • (optional) graph processing from adjacency list to adjacency matrix seems an evidence showing human consumes matter and energy to increase its order.

p3:

  • multi-graph (variable number of types/relations) vs multi-relational graph (fixed number of types/relations)
  • Heterogeneous graph (inner edges important than inter edges), multipartite graph (no inner edges, only inter edges), multiplex graph (inter edges important than inner edges)
  • attribute or feature

p4:

  • graph (abstract structure) and network (real-world data)
  • supervised (predict an output) and unsupervised (infer pattern)
  • node classification: predict the label of a node given a small number of labelled nodes (|V_train| « |V|)

p5:

  • applications: bot detection in a social network, function of proteins in the interactome, classify the topic based on links, etc
  • difference from a standard supervised learning: the assumption/bias of iid (independent and identically distributed) or no.
  • popular inductive bias used in graph learning: homophily (same attrubute with neighbors), structural equivalence (similar local structure → similar label), heterophily (e.g., gender).

p6:

  • supervised learning and semi-supervised learning, and GL (no iid assumption)
  • relation prediction: e.g., recommendation system, side-effect. Notice the requirement of inductive bias.

p7:

  • clustering and community detection
  • graph classification, regression, and clustering (to the audience: what is the general difference?)

p8:

  • iid assumption and why? → Li-Yang
  • Additional comment: Causal relation and correlation. ML is often consider the latter approach but actually we need to consider the former.
public/grl_readingmemo.txt · Last modified: 2024/04/04 16:12 by liang