Enhancing large language models for data analytics through domain-specific context creation
Liuska, Jonna (2024-04-05)
Liuska, Jonna
J. Liuska
05.04.2024
© 2024 Jonna Liuska. Ellei toisin mainita, uudelleenkäyttö on sallittu Creative Commons Attribution 4.0 International (CC-BY 4.0) -lisenssillä (https://creativecommons.org/licenses/by/4.0/). Uudelleenkäyttö on sallittua edellyttäen, että lähde mainitaan asianmukaisesti ja mahdolliset muutokset merkitään. Sellaisten osien käyttö tai jäljentäminen, jotka eivät ole tekijän tai tekijöiden omaisuutta, saattaa edellyttää lupaa suoraan asianomaisilta oikeudenhaltijoilta.
Julkaisun pysyvä osoite on
https://urn.fi/URN:NBN:fi:oulu-202404052578
https://urn.fi/URN:NBN:fi:oulu-202404052578
Tiivistelmä
This thesis was written to see the current situation in how the results and outputs of large language models (LLMs) can be enhanced by creating context. The two methods researched were prompting and retrieval augmented generation (RAG). These were chosen because prompting is seen as an easy-to-do solution and RAG is thought to decrease model hallucinations effectively. The thesis also covers how LLMs can be used for data analytics related tasks.
The results show that prompting is a bit more complex than usually thought of. Providing proper prompts was seen to reduce hallucinations, but then again, providing the prompts is not easy and requires planning and possibly multiple trial-and-error runs. Of the two methods RAG proved to be more effective for reducing hallucinations and providing better results. The big downside of RAG was found to be that it can get costly to implement and require experts to do it.
These results showed that currently there are effective methods for enhancing large language models, but further research is required on the topic.
The results show that prompting is a bit more complex than usually thought of. Providing proper prompts was seen to reduce hallucinations, but then again, providing the prompts is not easy and requires planning and possibly multiple trial-and-error runs. Of the two methods RAG proved to be more effective for reducing hallucinations and providing better results. The big downside of RAG was found to be that it can get costly to implement and require experts to do it.
These results showed that currently there are effective methods for enhancing large language models, but further research is required on the topic.
Kokoelmat
- Avoin saatavuus [37647]