نتایج جستجو

1

نتیجه یافت شد

مرتبط ترین ها

اعمال فیلتر

به روزترین ها

اعمال فیلتر

پربازدید ترین ها

اعمال فیلتر

پر دانلودترین‌ها

اعمال فیلتر

پر استنادترین‌ها

اعمال فیلتر

تعداد صفحات

1

انتقال به صفحه



فیلترها/جستجو در نتایج    

فیلترها

سال

بانک‌ها


گروه تخصصی


متن کامل


مرکز اطلاعات علمی SID1
اطلاعات دوره: 
  • سال: 

    2023
  • دوره: 

    11
  • شماره: 

    43
  • صفحات: 

    176-184
تعامل: 
  • استنادات: 

    0
  • بازدید: 

    0
  • دانلود: 

    0
چکیده: 

Text chunking is one of the basic tasks in natural language processing. Most proposed models in recent years were employed on chunking and other sequence labeling tasks simultaneously and they were mostly based on Recurrent Neural Networks (RNN) and Conditional Random Field (CRF). In this article, we use state-of-the-art transformer-based models in combination with CRF, Long Short-Term Memory (LSTM)-CRF as well as a simple dense layer to study the impact of different pre-trained models on the overall performance in text chunking. To this aim, we evaluate BERT, RoBERTa, Funnel Transformer, XLM, XLM-RoBERTa, BART, and GPT2 as candidates of contextualized models. Our experiments exhibit that all transformer-based models except GPT2 achieved close and high scores on text chunking. Due to the unique unidirectional architecture of GPT2, it shows a relatively poor performance on text chunking in comparison to other bidirectional transformer-based architectures. Our experiments also revealed that adding a LSTM layer to transformer-based models does not significantly improve the results since LSTM does not add additional features to assist the model to achieve more information from the input compared to the deep contextualized models.

آمار یکساله:   مرکز اطلاعات علمی Scientific Information Database (SID) - Trusted Source for Research and Academic Resources

بازدید 0

مرکز اطلاعات علمی Scientific Information Database (SID) - Trusted Source for Research and Academic Resourcesدانلود 0 مرکز اطلاعات علمی Scientific Information Database (SID) - Trusted Source for Research and Academic Resourcesاستناد 0 مرکز اطلاعات علمی Scientific Information Database (SID) - Trusted Source for Research and Academic Resourcesمرجع 0
litScript