全文获取类型
收费全文 | 146篇 |
免费 | 0篇 |
国内免费 | 25篇 |
专业分类
教育 | 50篇 |
科学研究 | 72篇 |
体育 | 2篇 |
综合类 | 7篇 |
文化理论 | 1篇 |
信息传播 | 39篇 |
出版年
2023年 | 2篇 |
2022年 | 4篇 |
2021年 | 3篇 |
2020年 | 2篇 |
2019年 | 6篇 |
2018年 | 3篇 |
2017年 | 4篇 |
2016年 | 1篇 |
2015年 | 5篇 |
2014年 | 13篇 |
2013年 | 16篇 |
2012年 | 11篇 |
2011年 | 19篇 |
2010年 | 8篇 |
2009年 | 15篇 |
2008年 | 9篇 |
2007年 | 18篇 |
2006年 | 14篇 |
2005年 | 5篇 |
2004年 | 6篇 |
2003年 | 3篇 |
2001年 | 1篇 |
2000年 | 1篇 |
1999年 | 1篇 |
1986年 | 1篇 |
排序方式: 共有171条查询结果,搜索用时 187 毫秒
71.
《Information processing & management》2022,59(3):102921
Efficient topic modeling is needed to support applications that aim at identifying main themes from a collection of documents. In the present paper, a reduced vector embedding representation and particle swarm optimization (PSO) are combined to develop a topic modeling strategy that is able to identify representative themes from a large collection of documents. Documents are encoded using a reduced, contextual vector embedding from a general-purpose pre-trained language model (sBERT). A modified PSO algorithm (pPSO) that tracks particle fitness on a dimension-by-dimension basis is then applied to these embeddings to create clusters of related documents. The proposed methodology is demonstrated on two datasets. The first dataset consists of posts from the online health forum r/Cancer and the second dataset is a standard benchmark for topic modeling which consists of a collection of messages posted to 20 different news groups. When compared to the state-of-the-art generative document models (i.e., ETM and NVDM), pPSO is able to produce interpretable clusters. The results indicate that pPSO is able to capture both common topics as well as emergent topics. Moreover, the topic coherence of pPSO is comparable to that of ETM and its topic diversity is comparable to NVDM. The assignment parity of pPSO on a document completion task exceeded 90% for the 20NewsGroups dataset. This rate drops to approximately 30% when pPSO is applied to the same Skip-Gram embedding derived from a limited, corpus-specific vocabulary which is used by ETM and NVDM. 相似文献
72.
基于迭代自组织数据聚类阈值的脉冲耦合神经网络的图像分割算法改进了传统脉冲耦合神经网络在图像分割中由于不恰当的参数选择而导致图像欠分割和过分割的问题.基于迭代自组织数据聚类阈值的脉冲耦合神经网络图像分割算法无需确定参数和循环次数,也不需要用特定原则确定循环结束的条件,只需利用图像中的每个像素点的灰度值进行聚类,然后利用改进的迭代自组织数据算法确定图像的初始聚类数目以及聚类中心,并以此作为脉冲耦合神经网络的最佳阈值,一次点火过程自动完成分割.实验结果表明,这种算法具有较好的分割结果和分割速度,提高了分割的准确性. 相似文献
73.
74.
Arabic is a widely spoken language but few mining tools have been developed to process Arabic text. This paper examines the crime domain in the Arabic language (unstructured text) using text mining techniques. The development and application of a Crime Profiling System (CPS) is presented. The system is able to extract meaningful information, in this case the type of crime, location and nationality, from Arabic language crime news reports. The system has two unique attributes; firstly, information extraction that depends on local grammar, and secondly, dictionaries that can be automatically generated. It is shown that the CPS improves the quality of the data through reduction where only meaningful information is retained. Moreover, the Self Organising Map (SOM) approach is adopted in order to perform the clustering of the crime reports, based on crime type. This clustering technique is improved because only refined data containing meaningful keywords extracted through the information extraction process are inputted into it, i.e. the data are cleansed by removing noise. The proposed system is validated through experiments using a corpus collated from different sources; it was not used during system development. Precision, recall and F-measure are used to evaluate the performance of the proposed information extraction approach. Also, comparisons are conducted with other systems. In order to evaluate the clustering performance, three parameters are used: data size, loading time and quantization error. 相似文献
75.
张英武 《鞍山师范学院学报》2013,(4):46-49
为了解决K-means算法中对于初值的敏感,提出了一种基于粒子群的改进的K-means聚类算法(IPSOFCM).在K-means算法中引入粒子群算法,可有效提高算法的全局搜索能力,有助于粒子更容易跳出局部束缚.实验结果证明,IPSOFCM算法聚类准确度高,稳定性好. 相似文献
76.
结合粗糙集理论,利用像素邻域的空间信息,可以构造图像色彩分布的上下近似以及量化粗糙性表示,据此提出一种基于量化粗糙信息的改进的图像分割方法,该方法使用局部量化粗糙度和待定算子来更新FCM算法中的隶属度函数。通过对比传统的模糊C-均值(FCM)聚类分割算法,证明该方法大大降低了时间复杂度,且具有良好的分割效果。 相似文献
77.
Ismail Sengor Altingovde Özlem Nurcan Subakan Özgür Ulusoy 《Information processing & management》2013
In-memory nearest neighbor computation is a typical collaborative filtering approach for high recommendation accuracy. However, this approach is not scalable given the huge number of customers and items in typical commercial applications. Cluster-based collaborative filtering techniques can be a remedy for the efficiency problem, but they usually provide relatively lower accuracy figures, since they may become over-generalized and produce less-personalized recommendations. Our research explores an individualistic strategy which initially clusters the users and then exploits the members within clusters, but not just the cluster representatives, during the recommendation generation stage. We provide an efficient implementation of this strategy by adapting a specifically tailored cluster-skipping inverted index structure. Experimental results reveal that the individualistic strategy with the cluster-skipping index is a good compromise that yields high accuracy and reasonable scalability figures. 相似文献
78.
In this paper, a Generalized Cluster Centroid based Classifier (GCCC) and its variants for text categorization are proposed by utilizing a clustering algorithm to integrate two well-known classifiers, i.e., the K-nearest-neighbor (KNN) classifier and the Rocchio classifier. KNN, a lazy learning method, suffers from inefficiency in online categorization while achieving remarkable effectiveness. Rocchio, which has efficient categorization performance, fails to obtain an expressive categorization model due to its inherent linear separability assumption. Our proposed method mainly focuses on two points: one point is that we use a clustering algorithm to strengthen the expressiveness of the Rocchio model; another one is that we employ the improved Rocchio model to speed up the categorization process of KNN. Extensive experiments conducted on both English and Chinese corpora show that GCCC and its variants have better categorization ability than some state-of-the-art classifiers, i.e., Rocchio, KNN and Support Vector Machine (SVM). 相似文献
79.
To address the inability of current ranking systems to support subtopic retrieval, two main post-processing techniques of search results have been investigated: clustering and diversification. In this paper we present a comparative study of their performance, using a set of complementary evaluation measures that can be applied to both partitions and ranked lists, and two specialized test collections focusing on broad and ambiguous queries, respectively. The main finding of our experiments is that diversification of top hits is more useful for quick coverage of distinct subtopics whereas clustering is better for full retrieval of single subtopics, with a better balance in performance achieved through generating multiple subsets of diverse search results. We also found that there is little scope for improvement over the search engine baseline unless we are interested in strict full-subtopic retrieval, and that search results clustering methods do not perform well on queries with low divergence subtopics, mainly due to the difficulty of generating discriminative cluster labels. 相似文献
80.
基于概念格的数字图书馆用户市场细分* ——数字图书馆用户的概念聚类分析 总被引:1,自引:0,他引:1
以概念格理论为基础,借助营销学中市场细分的变量,通过概念聚类,用形式概念分析的方法对数字图书馆用户进行市场细分。对在数字图书馆用户细分中突破传统统计口径、建立可伸缩的细分机制进行探索。 相似文献