Conference Paper/Proceeding/Abstract 79 views
GlossGPT: GPT for Word Sense Disambiguation using Few-shot Chain-of-Thought Prompting
The 8th International Conference on Emerging Data and Industry (EDI40)
Swansea University Authors:
Deshan Sumanathilaka , Nicholas Micallef
, Julian Hough
Abstract
Lexical ambiguity is a major challenge in computational linguistic tasks, as limitations in proper sense identification lead to inefficient translation and question answering. General-purpose Large Language Models (LLMs) are commonly utilized for Natural Language Processing (NLP) tasks. However, uti...
Published in: | The 8th International Conference on Emerging Data and Industry (EDI40) |
---|---|
Published: |
Patras, Greece
Elsevier Science
|
URI: | https://cronfa.swan.ac.uk/Record/cronfa68937 |
first_indexed |
2025-02-21T11:52:11Z |
---|---|
last_indexed |
2025-03-12T05:35:36Z |
id |
cronfa68937 |
recordtype |
SURis |
fullrecord |
<?xml version="1.0"?><rfc1807><datestamp>2025-03-11T13:23:55.3416739</datestamp><bib-version>v2</bib-version><id>68937</id><entry>2025-02-21</entry><title>GlossGPT: GPT for Word Sense Disambiguation using Few-shot Chain-of-Thought Prompting</title><swanseaauthors><author><sid>2fe44f0c1e7d845dc21bb6b00d5b2085</sid><ORCID>0009-0005-8933-6559</ORCID><firstname>Deshan</firstname><surname>Sumanathilaka</surname><name>Deshan Sumanathilaka</name><active>true</active><ethesisStudent>false</ethesisStudent></author><author><sid>1cc4c84582d665b7ee08fb16f5454671</sid><ORCID>0000-0002-2683-8042</ORCID><firstname>Nicholas</firstname><surname>Micallef</surname><name>Nicholas Micallef</name><active>true</active><ethesisStudent>false</ethesisStudent></author><author><sid>082d773ae261d2bbf49434dd2608ab40</sid><ORCID>0000-0002-4345-6759</ORCID><firstname>Julian</firstname><surname>Hough</surname><name>Julian Hough</name><active>true</active><ethesisStudent>false</ethesisStudent></author></swanseaauthors><date>2025-02-21</date><deptcode>MACS</deptcode><abstract>Lexical ambiguity is a major challenge in computational linguistic tasks, as limitations in proper sense identification lead to inefficient translation and question answering. General-purpose Large Language Models (LLMs) are commonly utilized for Natural Language Processing (NLP) tasks. However, utilizing general-purpose LLMs for specific tasks has been challenging, and fine-tuning has become a critical requirement for task specification. In this work, we craft advanced prompts with different contextual parameters to guide the model's inference towards accurate sense prediction to handle Word Sense Disambiguation (WSD). We present a few-shot Chain of Thought (COT) prompt-based technique using GPT-4-Turbo with knowledgebase as a retriever that does not require fine-tuning the model for WSD tasks and sense definitions are supported by synonyms to broaden the lexical meaning. Our approach achieves comparable performance on the SemEval and Senseval datasets. More importantly, we set a new state-of-the-art performance with the few-shot FEWS dataset, breaking through the 90% F1 score barrier.</abstract><type>Conference Paper/Proceeding/Abstract</type><journal>The 8th International Conference on Emerging Data and Industry (EDI40)</journal><volume/><journalNumber/><paginationStart/><paginationEnd/><publisher>Elsevier Science</publisher><placeOfPublication>Patras, Greece</placeOfPublication><isbnPrint/><isbnElectronic/><issnPrint/><issnElectronic/><keywords>Word Sense Disambiguation, Knowledge Base Retrieval, Large Language Models, GPT-4-Turbo, Chain of Thought Prompting</keywords><publishedDay>0</publishedDay><publishedMonth>0</publishedMonth><publishedYear>0</publishedYear><publishedDate>0001-01-01</publishedDate><doi/><url/><notes/><college>COLLEGE NANME</college><department>Mathematics and Computer Science School</department><CollegeCode>COLLEGE CODE</CollegeCode><DepartmentCode>MACS</DepartmentCode><institution>Swansea University</institution><apcterm>Not Required</apcterm><funders/><projectreference/><lastEdited>2025-03-11T13:23:55.3416739</lastEdited><Created>2025-02-21T11:46:03.2490240</Created><path><level id="1">Faculty of Science and Engineering</level><level id="2">School of Mathematics and Computer Science - Computer Science</level></path><authors><author><firstname>Deshan</firstname><surname>Sumanathilaka</surname><orcid>0009-0005-8933-6559</orcid><order>1</order></author><author><firstname>Nicholas</firstname><surname>Micallef</surname><orcid>0000-0002-2683-8042</orcid><order>2</order></author><author><firstname>Julian</firstname><surname>Hough</surname><orcid>0000-0002-4345-6759</orcid><order>3</order></author></authors><documents/><OutputDurs/></rfc1807> |
spelling |
2025-03-11T13:23:55.3416739 v2 68937 2025-02-21 GlossGPT: GPT for Word Sense Disambiguation using Few-shot Chain-of-Thought Prompting 2fe44f0c1e7d845dc21bb6b00d5b2085 0009-0005-8933-6559 Deshan Sumanathilaka Deshan Sumanathilaka true false 1cc4c84582d665b7ee08fb16f5454671 0000-0002-2683-8042 Nicholas Micallef Nicholas Micallef true false 082d773ae261d2bbf49434dd2608ab40 0000-0002-4345-6759 Julian Hough Julian Hough true false 2025-02-21 MACS Lexical ambiguity is a major challenge in computational linguistic tasks, as limitations in proper sense identification lead to inefficient translation and question answering. General-purpose Large Language Models (LLMs) are commonly utilized for Natural Language Processing (NLP) tasks. However, utilizing general-purpose LLMs for specific tasks has been challenging, and fine-tuning has become a critical requirement for task specification. In this work, we craft advanced prompts with different contextual parameters to guide the model's inference towards accurate sense prediction to handle Word Sense Disambiguation (WSD). We present a few-shot Chain of Thought (COT) prompt-based technique using GPT-4-Turbo with knowledgebase as a retriever that does not require fine-tuning the model for WSD tasks and sense definitions are supported by synonyms to broaden the lexical meaning. Our approach achieves comparable performance on the SemEval and Senseval datasets. More importantly, we set a new state-of-the-art performance with the few-shot FEWS dataset, breaking through the 90% F1 score barrier. Conference Paper/Proceeding/Abstract The 8th International Conference on Emerging Data and Industry (EDI40) Elsevier Science Patras, Greece Word Sense Disambiguation, Knowledge Base Retrieval, Large Language Models, GPT-4-Turbo, Chain of Thought Prompting 0 0 0 0001-01-01 COLLEGE NANME Mathematics and Computer Science School COLLEGE CODE MACS Swansea University Not Required 2025-03-11T13:23:55.3416739 2025-02-21T11:46:03.2490240 Faculty of Science and Engineering School of Mathematics and Computer Science - Computer Science Deshan Sumanathilaka 0009-0005-8933-6559 1 Nicholas Micallef 0000-0002-2683-8042 2 Julian Hough 0000-0002-4345-6759 3 |
title |
GlossGPT: GPT for Word Sense Disambiguation using Few-shot Chain-of-Thought Prompting |
spellingShingle |
GlossGPT: GPT for Word Sense Disambiguation using Few-shot Chain-of-Thought Prompting Deshan Sumanathilaka Nicholas Micallef Julian Hough |
title_short |
GlossGPT: GPT for Word Sense Disambiguation using Few-shot Chain-of-Thought Prompting |
title_full |
GlossGPT: GPT for Word Sense Disambiguation using Few-shot Chain-of-Thought Prompting |
title_fullStr |
GlossGPT: GPT for Word Sense Disambiguation using Few-shot Chain-of-Thought Prompting |
title_full_unstemmed |
GlossGPT: GPT for Word Sense Disambiguation using Few-shot Chain-of-Thought Prompting |
title_sort |
GlossGPT: GPT for Word Sense Disambiguation using Few-shot Chain-of-Thought Prompting |
author_id_str_mv |
2fe44f0c1e7d845dc21bb6b00d5b2085 1cc4c84582d665b7ee08fb16f5454671 082d773ae261d2bbf49434dd2608ab40 |
author_id_fullname_str_mv |
2fe44f0c1e7d845dc21bb6b00d5b2085_***_Deshan Sumanathilaka 1cc4c84582d665b7ee08fb16f5454671_***_Nicholas Micallef 082d773ae261d2bbf49434dd2608ab40_***_Julian Hough |
author |
Deshan Sumanathilaka Nicholas Micallef Julian Hough |
author2 |
Deshan Sumanathilaka Nicholas Micallef Julian Hough |
format |
Conference Paper/Proceeding/Abstract |
container_title |
The 8th International Conference on Emerging Data and Industry (EDI40) |
institution |
Swansea University |
publisher |
Elsevier Science |
college_str |
Faculty of Science and Engineering |
hierarchytype |
|
hierarchy_top_id |
facultyofscienceandengineering |
hierarchy_top_title |
Faculty of Science and Engineering |
hierarchy_parent_id |
facultyofscienceandengineering |
hierarchy_parent_title |
Faculty of Science and Engineering |
department_str |
School of Mathematics and Computer Science - Computer Science{{{_:::_}}}Faculty of Science and Engineering{{{_:::_}}}School of Mathematics and Computer Science - Computer Science |
document_store_str |
0 |
active_str |
0 |
description |
Lexical ambiguity is a major challenge in computational linguistic tasks, as limitations in proper sense identification lead to inefficient translation and question answering. General-purpose Large Language Models (LLMs) are commonly utilized for Natural Language Processing (NLP) tasks. However, utilizing general-purpose LLMs for specific tasks has been challenging, and fine-tuning has become a critical requirement for task specification. In this work, we craft advanced prompts with different contextual parameters to guide the model's inference towards accurate sense prediction to handle Word Sense Disambiguation (WSD). We present a few-shot Chain of Thought (COT) prompt-based technique using GPT-4-Turbo with knowledgebase as a retriever that does not require fine-tuning the model for WSD tasks and sense definitions are supported by synonyms to broaden the lexical meaning. Our approach achieves comparable performance on the SemEval and Senseval datasets. More importantly, we set a new state-of-the-art performance with the few-shot FEWS dataset, breaking through the 90% F1 score barrier. |
published_date |
0001-01-01T10:24:57Z |
_version_ |
1827108340363165696 |
score |
11.055543 |