Looking Back to Look Forward — Long before Swin Transformer[1], Cluster-GCN uses inter-cluster links inside a mini-batch, which facilitates the information flow between clusters. TBC [1] Liu, Ze, et al. “Swin transformer: Hierarchical vision transformer using shifted windows.” Proceedings of the IEEE/CVF international conference on computer vision. 2021. [2] Chiang, Wei-Lin, et al. “Cluster-gcn: An efficient algorithm for training deep and large graph convolutional networks.” Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining. 2019.