首页> 外文会议>NAFOSTED Conference on Information and Computer Science >Vietnamese Facebook Posts Classification using Fine-Tuning BERT
【24h】

Vietnamese Facebook Posts Classification using Fine-Tuning BERT

机译:越南Facebook帖子使用微调伯特分类

获取原文

摘要

With the development of social networks in the age of information technology explosion, the classification of social news plays an important role in detecting the hot topics being discussed on social networks over a period of time. In this paper, we present a new model for effective Facebook's posts classification and a new dataset which is labeled for the corresponding subject. The dataset consists of 5191 Facebook user's public posts, which is divided into 3 subsets: training, validation and testing data sets. Then, we explore the effectiveness of fine-tuning BERT model with three truncation methods compared with other machine learning algorithms on our dataset. Experimental results show that the fine-tune BERT models outperform other approaches. The fine-tune BERT with “head + tail” truncation methods achieves the best scores with 84.31% of Precision, 84.12% of Recall and 84.15% of F1-score.
机译:随着社会网络的发展,在信息技术爆炸时,社会新闻的分类在检测到在一段时间内检测在社交网络上讨论的热门话题中起着重要作用。在本文中,我们为有效的Facebook的帖子分类和一个标有相应主题的新数据集提供了一个新模型。 DataSet由5191 Facebook用户的公共帖子组成,该帖子分为3个子集:培训,验证和测试数据集。然后,我们探讨了微调BERT模型与三个截断方法的有效性与我们数据集上的其他机器学习算法相比。实验结果表明,微调BERT模型优于其他方法。具有“头部+尾”截断方法的微调伯特达到了最佳分数,精度的84.31%,召回的84.12%和84.15%的F1分数。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号