179 research outputs found

    黃仲則寓京期間感遇詩之研究

    Full text link
    有清一代,文風大盛。不論詩詞歌賦,抑是經史樸學,無不興盛。及至康、乾二朝,大家漸出。其中尤以詩壇之中名家輩出,然而清代的詩人,往往受唐宋名家所迷,困於前代詩人的範曙中,劉大杰曾謂: 清代詩人,喜言宗派 。[1] 清代詩派,大要言之,有尊唐及宗宋二派主流,而其中,尤以四家詩說影響至鉅。所謂四家詩說是指王士禎的神韻說、沈德潛的格調說、袁枚的性靈說及翁方綱的肌理說。在此四家詩說之下,康乾詩人,鮮有不受影響。而芸芸詩人中,黃景仁可說是少數能獨樹一幟的詩人。 今觀其詩作,既沒有神韻說的空洞,又沒有格調說的泥古不化,亦沒有肌種說的堆集煩瑣。可以說,他的詩歌,較近於袁枚的性重說,此或因黃袁二人為忘年之交,彼此有詩文來往,稱頌對方,故受到性靈說所黨陶的緣故。[2] 惟其詩雖近袁子才的性靈說, 但卻沒有半點袁詩通俗淺白的弊病。 若觀乎《兩當軒集》中諸詩,字字出自肺肺,莫不充溢其個人深厚的感情。正因如此,是以仲則在生前後之文,均對其詩有極高之評價,清代翁方綱云: 然而其詩尚沈鬱青壯,鏗鏘出金石,試摘其一二語,可通風雪而泣鬼神。[3] 清包世臣亦謂: 聲稱噪一時,乾隆六十年間,論詩者推為第一。[4] 清吳嵩梁於《石溪勛講話》中,亦有提及仲則的詩: 仲則詩無奇不有,無妙不臻,如仙人張樂,音外有音;名將用兵,法外有法。[5] 清汪防亦云: 吾鄉黃仲則先生,以詩鳴乾隆中葉。[6] 又,清代邱煒云: 乾隆才子黃仲刻,詩名遠播。[7] 近人伍合曰: 我們知道,作詩要下一番苦工夫,起碼非有幾十年的磨練,詩才會好,所以古人說, 晚節漸於詩律細 ,那一點也不錯,至於天賦的,那真是絕無僅有,往上數,李太白可以算一個,往下數,那只有黃景仁了。[8] 近人君山亦云: 黃仲則在世的日子雖然不長,死時僅得三十五歲,可是,在乾隆年間甚至乾隆以後,他卻是影響清代詩壇最大的一個詩人。[9] 自以上資料可見,黃仲則的詩作在其生前及死後,均得到肯定的評價。而在他芸芸詩作中,仲則的感遇詩最為世人所重。此蓋因其身世坎呵,生活艱苦,一生奔波,卻屢不得志,是故其詩作,情感特真,感慨特深。惟論者往往忽視仲則一生最感傷悲,精神所受痛苦最深切的時期,應當是他久滯京師時,是故仲則此時期的感遇詩,亦最能作為他同類詩作的代表。故此一時期的感遇詩,是極為值得研究的。 本文的研究範圍,當以仲則滯留京師至其歿於山西解州為止。考仲則一生,自乾隆四十年冬入京後,至乾隆四十八年病死止,凡三離京師,計有四十五年遊山東,客於學政程世淳幕中;次為四十六年遊西安,往訪峽西巡撫畢況;而最後一次為四十八年三月,因債主所迫,故抱病出都,至解州病歿。除最後一次以外,餘二次出京,為時皆甚短,故可說仲則寓京時間大約前後只有九年。因此本文研究之範團亦以此段時期為主。據《爾當軒集》中所編,仲則於此段時期約有詩作三百七十多首。本文所研究之對象為其感遇詩,故凡詩題中有感遇等字眼者,自是研究對象;另外尚有其他詩歌,其題雖不涉及感遇等字眼,惟若真內容有抒感言懷者,皆視之為感遇詩,一并研究。而於三百七十多首詩歌中,合乎於感遇題材之詩歌共有一百零九首,是以研究將集中於此一百零九首詩歌,惟若有其他可資佐證之詩歌,雖非於寓京期間所作,本文亦會加以引用。而文中黃仲則在京師時的感遇詩作,將一律簡稱為感遇詩,若遇有寓京以外的感遇詩,本文將加以注明。 本文共分八個部分。除前言後語與及文末注釋及參考書籍部分外,主要內容共分為四個部分:分別為(i) 窮 ; (ii) 貧 ; (iii) 病 ; (iv) 孤 。本文希望通過這四部分,對黃仲則寓京期間的感遇詩的感遇緣起作一研究,非去有所補足,只望可資作參考之用

    ELCoRec: Enhance Language Understanding with Co-Propagation of Numerical and Categorical Features for Recommendation

    Full text link
    Large language models have been flourishing in the natural language processing (NLP) domain, and their potential for recommendation has been paid much attention to. Despite the intelligence shown by the recommendation-oriented finetuned models, LLMs struggle to fully understand the user behavior patterns due to their innate weakness in interpreting numerical features and the overhead for long context, where the temporal relations among user behaviors, subtle quantitative signals among different ratings, and various side features of items are not well explored. Existing works only fine-tune a sole LLM on given text data without introducing that important information to it, leaving these problems unsolved. In this paper, we propose ELCoRec to Enhance Language understanding with CoPropagation of numerical and categorical features for Recommendation. Concretely, we propose to inject the preference understanding capability into LLM via a GAT expert model where the user preference is better encoded by parallelly propagating the temporal relations, and rating signals as well as various side information of historical items. The parallel propagation mechanism could stabilize heterogeneous features and offer an informative user preference encoding, which is then injected into the language models via soft prompting at the cost of a single token embedding. To further obtain the user's recent interests, we proposed a novel Recent interaction Augmented Prompt (RAP) template. Experiment results over three datasets against strong baselines validate the effectiveness of ELCoRec. The code is available at https://anonymous.4open.science/r/CIKM_Code_Repo-E6F5/README.md

    AGADIR: Towards Array-Geometry Agnostic Directional Speech Recognition

    Full text link
    Wearable devices like smart glasses are approaching the compute capability to seamlessly generate real-time closed captions for live conversations. We build on our recently introduced directional Automatic Speech Recognition (ASR) for smart glasses that have microphone arrays, which fuses multi-channel ASR with serialized output training, for wearer/conversation-partner disambiguation as well as suppression of cross-talk speech from non-target directions and noise. When ASR work is part of a broader system-development process, one may be faced with changes to microphone geometries as system development progresses. This paper aims to make multi-channel ASR insensitive to limited variations of microphone-array geometry. We show that a model trained on multiple similar geometries is largely agnostic and generalizes well to new geometries, as long as they are not too different. Furthermore, training the model this way improves accuracy for seen geometries by 15 to 28\% relative. Lastly, we refine the beamforming by a novel Non-Linearly Constrained Minimum Variance criterion.Comment: Accepted to ICASSP 202

    M-scan: A Multi-Scenario Causal-driven Adaptive Network for Recommendation

    Full text link
    We primarily focus on the field of multi-scenario recommendation, which poses a significant challenge in effectively leveraging data from different scenarios to enhance predictions in scenarios with limited data. Current mainstream efforts mainly center around innovative model network architectures, with the aim of enabling the network to implicitly acquire knowledge from diverse scenarios. However, the uncertainty of implicit learning in networks arises from the absence of explicit modeling, leading to not only difficulty in training but also incomplete user representation and suboptimal performance. Furthermore, through causal graph analysis, we have discovered that the scenario itself directly influences click behavior, yet existing approaches directly incorporate data from other scenarios during the training of the current scenario, leading to prediction biases when they directly utilize click behaviors from other scenarios to train models. To address these problems, we propose the Multi-Scenario Causal-driven Adaptive Network M-scan). This model incorporates a Scenario-Aware Co-Attention mechanism that explicitly extracts user interests from other scenarios that align with the current scenario. Additionally, it employs a Scenario Bias Eliminator module utilizing causal counterfactual inference to mitigate biases introduced by data from other scenarios. Extensive experiments on two public datasets demonstrate the efficacy of our M-scan compared to the existing baseline models.Comment: This paper has been accepted by WWW'2

    Towards Efficient and Effective Unlearning of Large Language Models for Recommendation

    Full text link
    The significant advancements in large language models (LLMs) give rise to a promising research direction, i.e., leveraging LLMs as recommenders (LLMRec). The efficacy of LLMRec arises from the open-world knowledge and reasoning capabilities inherent in LLMs. LLMRec acquires the recommendation capabilities through instruction tuning based on user interaction data. However, in order to protect user privacy and optimize utility, it is also crucial for LLMRec to intentionally forget specific user data, which is generally referred to as recommendation unlearning. In the era of LLMs, recommendation unlearning poses new challenges for LLMRec in terms of \textit{inefficiency} and \textit{ineffectiveness}. Existing unlearning methods require updating billions of parameters in LLMRec, which is costly and time-consuming. Besides, they always impact the model utility during the unlearning process. To this end, we propose \textbf{E2URec}, the first \underline{E}fficient and \underline{E}ffective \underline{U}nlearning method for LLM\underline{Rec}. Our proposed E2URec enhances the unlearning efficiency by updating only a few additional LoRA parameters, and improves the unlearning effectiveness by employing a teacher-student framework, where we maintain multiple teacher networks to guide the unlearning process. Extensive experiments show that E2URec outperforms state-of-the-art baselines on two real-world datasets. Specifically, E2URec can efficiently forget specific data without affecting recommendation performance. The source code is at \url{https://github.com/justarter/E2URec}.Comment: Accepted by Frontier of Computer Scienc

    Large Language Models Make Sample-Efficient Recommender Systems

    Full text link
    Large language models (LLMs) have achieved remarkable progress in the field of natural language processing (NLP), demonstrating remarkable abilities in producing text that resembles human language for various tasks. This opens up new opportunities for employing them in recommender systems (RSs). In this paper, we specifically examine the sample efficiency of LLM-enhanced recommender systems, which pertains to the model's capacity to attain superior performance with a limited quantity of training data. Conventional recommendation models (CRMs) often need a large amount of training data because of the sparsity of features and interactions. Hence, we propose and verify our core viewpoint: Large Language Models Make Sample-Efficient Recommender Systems. We propose a simple yet effective framework (i.e., Laser) to validate the viewpoint from two aspects: (1) LLMs themselves are sample-efficient recommenders; and (2) LLMs, as feature generators and encoders, make CRMs more sample-efficient. Extensive experiments on two public datasets show that Laser requires only a small fraction of training samples to match or even surpass CRMs that are trained on the entire training set, demonstrating superior sample efficiency.Comment: Accepted by Frontier of Computer Scienc

    ReLLa: Retrieval-enhanced Large Language Models for Lifelong Sequential Behavior Comprehension in Recommendation

    Full text link
    With large language models (LLMs) achieving remarkable breakthroughs in natural language processing (NLP) domains, LLM-enhanced recommender systems have received much attention and have been actively explored currently. In this paper, we focus on adapting and empowering a pure large language model for zero-shot and few-shot recommendation tasks. First and foremost, we identify and formulate the lifelong sequential behavior incomprehension problem for LLMs in recommendation domains, i.e., LLMs fail to extract useful information from a textual context of long user behavior sequence, even if the length of context is far from reaching the context limitation of LLMs. To address such an issue and improve the recommendation performance of LLMs, we propose a novel framework, namely Retrieval-enhanced Large Language models (ReLLa) for recommendation tasks in both zero-shot and few-shot settings. For zero-shot recommendation, we perform semantic user behavior retrieval (SUBR) to improve the data quality of testing samples, which greatly reduces the difficulty for LLMs to extract the essential knowledge from user behavior sequences. As for few-shot recommendation, we further design retrieval-enhanced instruction tuning (ReiT) by adopting SUBR as a data augmentation technique for training samples. Specifically, we develop a mixed training dataset consisting of both the original data samples and their retrieval-enhanced counterparts. We conduct extensive experiments on a real-world public dataset (i.e., MovieLens-1M) to demonstrate the superiority of ReLLa compared with existing baseline models, as well as its capability for lifelong sequential behavior comprehension.Comment: Under Revie

    ClickPrompt: CTR Models are Strong Prompt Generators for Adapting Language Models to CTR Prediction

    Full text link
    Click-through rate (CTR) prediction has become increasingly indispensable for various Internet applications. Traditional CTR models convert the multi-field categorical data into ID features via one-hot encoding, and extract the collaborative signals among features. Such a paradigm suffers from the problem of semantic information loss. Another line of research explores the potential of pretrained language models (PLMs) for CTR prediction by converting input data into textual sentences through hard prompt templates. Although semantic signals are preserved, they generally fail to capture the collaborative information (e.g., feature interactions, pure ID features), not to mention the unacceptable inference overhead brought by the huge model size. In this paper, we aim to model both the semantic knowledge and collaborative knowledge for accurate CTR estimation, and meanwhile address the inference inefficiency issue. To benefit from both worlds and close their gaps, we propose a novel model-agnostic framework (i.e., ClickPrompt), where we incorporate CTR models to generate interaction-aware soft prompts for PLMs. We design a prompt-augmented masked language modeling (PA-MLM) pretraining task, where PLM has to recover the masked tokens based on the language context, as well as the soft prompts generated by CTR model. The collaborative and semantic knowledge from ID and textual features would be explicitly aligned and interacted via the prompt interface. Then, we can either tune the CTR model with PLM for superior performance, or solely tune the CTR model without PLM for inference efficiency. Experiments on four real-world datasets validate the effectiveness of ClickPrompt compared with existing baselines.Comment: Accepted by WWW 202

    DisCo: Towards Harmonious Disentanglement and Collaboration between Tabular and Semantic Space for Recommendation

    Full text link
    Recommender systems play important roles in various applications such as e-commerce, social media, etc. Conventional recommendation methods usually model the collaborative signals within the tabular representation space. Despite the personalization modeling and the efficiency, the latent semantic dependencies are omitted. Methods that introduce semantics into recommendation then emerge, injecting knowledge from the semantic representation space where the general language understanding are compressed. However, existing semantic-enhanced recommendation methods focus on aligning the two spaces, during which the representations of the two spaces tend to get close while the unique patterns are discarded and not well explored. In this paper, we propose DisCo to Disentangle the unique patterns from the two representation spaces and Collaborate the two spaces for recommendation enhancement, where both the specificity and the consistency of the two spaces are captured. Concretely, we propose 1) a dual-side attentive network to capture the intra-domain patterns and the inter-domain patterns, 2) a sufficiency constraint to preserve the task-relevant information of each representation space and filter out the noise, and 3) a disentanglement constraint to avoid the model from discarding the unique information. These modules strike a balance between disentanglement and collaboration of the two representation spaces to produce informative pattern vectors, which could serve as extra features and be appended to arbitrary recommendation backbones for enhancement. Experiment results validate the superiority of our method against different models and the compatibility of DisCo over different backbones. Various ablation studies and efficiency analysis are also conducted to justify each model component

    Towards Open-World Recommendation with Knowledge Augmentation from Large Language Models

    Full text link
    Recommender systems play a vital role in various online services. However, the insulated nature of training and deploying separately within a specific domain limits their access to open-world knowledge. Recently, the emergence of large language models (LLMs) has shown promise in bridging this gap by encoding extensive world knowledge and demonstrating reasoning capability. Nevertheless, previous attempts to directly use LLMs as recommenders have not achieved satisfactory results. In this work, we propose an Open-World Knowledge Augmented Recommendation Framework with Large Language Models, dubbed KAR, to acquire two types of external knowledge from LLMs -- the reasoning knowledge on user preferences and the factual knowledge on items. We introduce factorization prompting to elicit accurate reasoning on user preferences. The generated reasoning and factual knowledge are effectively transformed and condensed into augmented vectors by a hybrid-expert adaptor in order to be compatible with the recommendation task. The obtained vectors can then be directly used to enhance the performance of any recommendation model. We also ensure efficient inference by preprocessing and prestoring the knowledge from the LLM. Extensive experiments show that KAR significantly outperforms the state-of-the-art baselines and is compatible with a wide range of recommendation algorithms. We deploy KAR to Huawei's news and music recommendation platforms and gain a 7\% and 1.7\% improvement in the online A/B test, respectively
    corecore