论文标题

suply-avae:基于对抗性变异自动编码器的属性推理模型

Infer-AVAE: An Attribute Inference Model Based on Adversarial Variational Autoencoder

论文作者

Zhou, Yadong, Ding, Zhihao, Liu, Xiaoming, Shen, Chao, Tong, Lingling, Guan, Xiaohong

论文摘要

性别和教育等用户属性在社交网络中面临严重的不完整。为了使这种有价值的数据可用于下游任务(例如用户分析和个性化建议),属性推理旨在根据观察到的数据推断用户缺少属性标签。最近,端到端的深层生成模型变量自动编码器(VAE)通过半监督的方式来处理该问题,显示出了有希望的性能。但是,当应用于属性推理时,VAE很容易遭受过度贴合和过度光滑的折磨。具体而言,使用多层感知器(MLP)实现的VAE只能重建输入数据,但无法推断丢失的零件。在使用趋势图神经网络(GNN)用作编码器的同时,GNNS从邻居中汇总了冗余信息,并生成难以区分的用户表示,这被称为过度光滑。在本文中,我们提出了一个基于\ textbf {a} dversarial \ textbf {vae}(subl-avae)的属性\ textbf {peastBf {celes} ence模型以应对这些问题。具体而言,为了克服过度光滑的,下avae在编码器中统一了MLP和GNN,分别学习了积极和负的潜在表示。同时,对对抗网络进行了训练,以区分这两种表示形式,而GNN经过培训,可以通过对抗性训练来汇总噪声更少,以促进更强大的表示。最后,为了缓解过度拟合,将共同信息约束作为解码器的正规化程序引入,以便它可以更好地利用表示中的辅助信息,并生成不受观察限制的输出。我们在4个现实世界的社交网络数据集上评估了我们的模型,实验结果表明,我们的模型的准确性平均优于7.0 $ \%$。

User attributes, such as gender and education, face severe incompleteness in social networks. In order to make this kind of valuable data usable for downstream tasks like user profiling and personalized recommendation, attribute inference aims to infer users' missing attribute labels based on observed data. Recently, variational autoencoder (VAE), an end-to-end deep generative model, has shown promising performance by handling the problem in a semi-supervised way. However, VAEs can easily suffer from over-fitting and over-smoothing when applied to attribute inference. To be specific, VAE implemented with multi-layer perceptron (MLP) can only reconstruct input data but fail in inferring missing parts. While using the trending graph neural networks (GNNs) as encoder has the problem that GNNs aggregate redundant information from neighborhood and generate indistinguishable user representations, which is known as over-smoothing. In this paper, we propose an attribute \textbf{Infer}ence model based on \textbf{A}dversarial \textbf{VAE} (Infer-AVAE) to cope with these issues. Specifically, to overcome over-smoothing, Infer-AVAE unifies MLP and GNNs in encoder to learn positive and negative latent representations respectively. Meanwhile, an adversarial network is trained to distinguish the two representations and GNNs are trained to aggregate less noise for more robust representations through adversarial training. Finally, to relieve over-fitting, mutual information constraint is introduced as a regularizer for decoder, so that it can make better use of auxiliary information in representations and generate outputs not limited by observations. We evaluate our model on 4 real-world social network datasets, experimental results demonstrate that our model averagely outperforms baselines by 7.0$\%$ in accuracy.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源