Abstract:
A multimodal recommendation (MMSAF)model that integrates high-order semantic enhancement with adaptive modal feature fusion is proposed. The model employs a graph convolutional neural network (GCNN)to perform high-order semantic enhancement, enabling the capture of deeper associations between users and items, thereby more accurately reflecting users’ complex preferences. The effectiveness and applicability of the model are validated. An adaptive modality fusion mechanism is introduced to dynamically adjust the weights of modal features based on their relative importance in different contexts, enabling flexible adaptation to diverse user preferences. Experimental results demonstrate that MMSAF significantly outperforms existing mainstream methods across multiple benchmark datasets in terms of recommendation accuracy and generalization capability.