政大機構典藏-National Chengchi University Institutional Repository(NCCUR):Item 140.119/147031
English  |  正體中文  |  简体中文  |  Post-Print筆數 : 27 |  全文笔数/总笔数 : 113318/144297 (79%)
造访人次 : 51093419      在线人数 : 731
RC Version 6.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
搜寻范围 查询小技巧:
  • 您可在西文检索词汇前后加上"双引号",以获取较精准的检索结果
  • 若欲以作者姓名搜寻,建议至进阶搜寻限定作者字段,可获得较完整数据
  • 进阶搜寻
    政大機構典藏 > 資訊學院 > 資訊科學系 > 學位論文 >  Item 140.119/147031


    请使用永久网址来引用或连结此文件: https://nccur.lib.nccu.edu.tw/handle/140.119/147031


    题名: 基於高階相鄰關係與最短路徑之圖表示法學習
    Exploring High-order Proximity and Shortest-path Walking for Graph Representation Learning
    作者: 林柏宇
    Lin, Bo-Yu
    贡献者: 蔡銘峰
    林柏宇
    Lin, Bo-Yu
    关键词: 圖學習表示法
    最短路徑
    鏈結預測
    Graph Representation Learning
    Shortest Path
    Link Prediction
    日期: 2023
    上传时间: 2023-09-01 15:24:12 (UTC+8)
    摘要: 圖學習表示法作為一種重要的圖分析方法,旨在將圖中的節點和邊映射到低維度的向量空間,以更精確地捕捉其結構和關係特徵。在這一框架中,高階相鄰關係扮演著關鍵角色。相對於傳統的低階表示法,高階相鄰關係提供了更深入的分析和處理能力。 它能夠捕捉到節點之間的複雜連接關係,包括間接連接和共同鄰居等。此外,高階相 鄰關係在圖的結構中具有重要性,如三角形關係和子圖模式等。通過學習這些關係,模型能夠更準確地理解圖的整體特徵。因此,探索高階相鄰關係對於進一步提升圖學習表示法的性能和能力至關重要。
    本研究旨在引入「最短路徑」作為主要概念,以改善圖學習表示法對高階相鄰關係的學習能力。通過利用最短路徑可達到 k 步為 k 階鄰 居的特性,我們希望表示法能更加直觀且準確地捕捉高階相鄰關係。此外,我們進一步探索以最短路徑長的分佈來動態決定高階鄰居的取值範圍。 這種方法使得圖學習表示法能夠自動調整參數,而無需耗費大量時間和計算資源。透過基於最短路徑長的分佈,我們能夠更有效地決定高階鄰居的範圍,從而提高圖學習表示法的性能和效率。
    透過以上方法,本研究希望為圖學習表示法在學習高階相鄰關係方面提供一個新的視角和改進策略,期待過程中的實驗結果與觀察發現能為未來相關的研究提供幫助。
    Graph representation learning is a critical technique in graph analysis that strives to project nodes and edges of a graph into a compressed vector space, thereby better grasping structural and relational aspects. Central to this are higher-order neighboring relationships. These relationships, unlike traditional lower-order ones, offer enhanced analysis and processing potential. They excel at detecting nuanced connections between nodes, such as indirect ties and shared neighbors. Importantly, these higher-order relation- ships highlight specific patterns in the graph structure, like triangle relation- ships and subgraph designs. By mastering these, models can more thoroughly comprehend the overarching graph features. Hence, examining higher-order neighboring relationships is essential for refining the efficacy of graph representation learning.
    In this research, we propose the ”shortest paths” principle to boost the learning capacity of graph representation concerning higher-order neighbor- ing relationships. By harnessing that the shortest paths of up to k steps produce kth-order neighbors, we aim for a more precise portrayal of these relationships. Additionally, we explore the adaptive determination of the span of higher-order neighbors using the spread of shortest path lengths. Such a strategy enables graph representation learning to self-regulate parameters with- out excessive resource expenditure. Leveraging the spread of shortest path lengths helps swiftly determine the range of higher-order neighbors, hence enhancing graph representation learning’s effectiveness.
    With these methodologies, our research offers a novel viewpoint and enhancement approach for graph representation learning, focusing on mastering higher-order neighboring relationships. The anticipated findings from this study will likely benefit subsequent studies in this domain.
    參考文獻: 1. Yoshua Bengio, Aaron Courville, and Pascal Vincent. Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence, 35(8):1798–1828, 2013.
    2. Bryan Perozzi, Rami Al-Rfou, and Steven Skiena. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD international con- ference on Knowledge discovery and data mining, pages 701–710, 2014.
    3. Aditya Grover and Jure Leskovec.node2vec:Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining, pages 855–864, 2016.
    4. XinRong.word2vec parameter learning explained.arXiv preprint arXiv:1411.2738, 2014.
    5. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017.
    6. Lisa Ehrlinger and Wolfram Wo ̈ß. Towards a definition of knowledge graphs. SE- MANTiCS (Posters, Demos, SuCCESS), 48(1-4):2, 2016.
    7. Jheng-Hong Yang, Chih-Ming Chen, Chuan-Ju Wang, and Ming-Feng Tsai. Hop- rec: high-order proximity for implicit recommendation. In Proceedings of the 12th ACM conference on recommender systems, pages 140–144, 2018
    8. Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, 33:22118– 22133, 2020.
    9. Jian Tang, Meng Qu, Mingzhe Wang, Ming Zhang, Jun Yan, and Qiaozhu Mei. Line: Large-scale information network embedding. In Proceedings of the 24th international conference on world wide web, pages 1067–1077, 2015.
    10. Gueorgi Kossinets. Effects of missing data in social networks. Social networks, 28(3):247–268, 2006.
    11. Albert-La ́szlo ́ Baraba ́si and Re ́ka Albert. Emergence of scaling in random networks. science, 286(5439):509–512, 1999.
    12.Lada A Adamic and Eytan Adar.Friends and neighbors on the web.Social networks, 25(3):211–230, 2003.
    13. Yehuda Koren, Robert Bell, and Chris Volinsky. Matrix factorization techniques for recommender systems. Computer, 42(8):30–37, 2009.
    14. Keiron O’Shea and Ryan Nash. An introduction to convolutional neural networks. arXiv preprint arXiv:1511.08458, 2015.
    15. Wojciech Zaremba, Ilya Sutskever, and Oriol Vinyals. Recurrent neural network regularization. arXiv preprint arXiv:1409.2329, 2014.
    16. Qiaoyu Tan, Xin Zhang, Ninghao Liu, Daochen Zha, Li Li, Rui Chen, Soo-Hyun Choi, and Xia Hu. Bring your own view: Graph neural networks for link predic- tion with personalized subgraph selection. In Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, pages 625–633, 2023.
    17.XiangnanHe,KuanDeng,XiangWang,YanLi,YongdongZhang,andMengWang. Lightgcn: Simplifying and powering graph convolution network for recommenda- tion. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, pages 639–648, 2020.
    18. Will Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. Advances in neural information processing systems, 30, 2017.
    19. Muhan Zhang and Yixin Chen. Link prediction based on graph neural networks. Advances in neural information processing systems, 31, 2018.
    20. Anh Viet Phan, Minh Le Nguyen, Yen Lam Hoang Nguyen, and Lam Thu Bui. Dgcnn: A convolutional neural network over large-scale labeled graphs. Neural Networks, 108:533–543, 2018.
    21. Zhitao Wang, Yong Zhou, Litao Hong, Yuanhang Zou, Hanjing Su, and Shouzhi Chen. Pairwise learning for neural link prediction. arXiv preprint arXiv:2112.02936, 2021.
    描述: 碩士
    國立政治大學
    資訊科學系
    110753109
    資料來源: http://thesis.lib.nccu.edu.tw/record/#G0110753109
    数据类型: thesis
    显示于类别:[資訊科學系] 學位論文

    文件中的档案:

    档案 描述 大小格式浏览次数
    310901.pdf2085KbAdobe PDF20检视/开启


    在政大典藏中所有的数据项都受到原著作权保护.


    社群 sharing

    著作權政策宣告 Copyright Announcement
    1.本網站之數位內容為國立政治大學所收錄之機構典藏,無償提供學術研究與公眾教育等公益性使用,惟仍請適度,合理使用本網站之內容,以尊重著作權人之權益。商業上之利用,則請先取得著作權人之授權。
    The digital content of this website is part of National Chengchi University Institutional Repository. It provides free access to academic research and public education for non-commercial use. Please utilize it in a proper and reasonable manner and respect the rights of copyright owners. For commercial use, please obtain authorization from the copyright owner in advance.

    2.本網站之製作,已盡力防止侵害著作權人之權益,如仍發現本網站之數位內容有侵害著作權人權益情事者,請權利人通知本網站維護人員(nccur@nccu.edu.tw),維護人員將立即採取移除該數位著作等補救措施。
    NCCU Institutional Repository is made to protect the interests of copyright owners. If you believe that any material on the website infringes copyright, please contact our staff(nccur@nccu.edu.tw). We will remove the work from the repository and investigate your claim.
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - 回馈