Evaluating the carbon footprint of code implementation
This project undertakes the evaluation of the carbon footprint of large language models (LLMs). Three models are evaluated – Meta’s LLaMA-2 (7-billion parameter configuration), Mistral (7-billion parameter configuration), and Google’s Gemma (2-billion and 7-billion parameter configurations). The emi...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/181172 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-181172 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1811722024-11-18T01:35:30Z Evaluating the carbon footprint of code implementation Tar, Sreeja Lim Wei Yang Bryan College of Computing and Data Science bryan.limwy@ntu.edu.sg Computer and Information Science Large Language Models (LLMs) Carbon footprint Fine-tuning Natural Language Processing (NLP) Emissions Energy consumption This project undertakes the evaluation of the carbon footprint of large language models (LLMs). Three models are evaluated – Meta’s LLaMA-2 (7-billion parameter configuration), Mistral (7-billion parameter configuration), and Google’s Gemma (2-billion and 7-billion parameter configurations). The emissions generated by these models in the fine-tuning phase are evaluated for three tasks – question answering, text summarisation, and sentiment analysis. This report also examines differences in emissions generated across GPUs. The project ultimately explores the impact of model optimisation on emissions and provides corresponding recommendations to reduce the carbon footprint of the selected models. Bachelor's degree 2024-11-18T01:35:06Z 2024-11-18T01:35:06Z 2024 Final Year Project (FYP) Tar, S. (2024). Evaluating the carbon footprint of code implementation. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181172 https://hdl.handle.net/10356/181172 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Large Language Models (LLMs) Carbon footprint Fine-tuning Natural Language Processing (NLP) Emissions Energy consumption |
spellingShingle |
Computer and Information Science Large Language Models (LLMs) Carbon footprint Fine-tuning Natural Language Processing (NLP) Emissions Energy consumption Tar, Sreeja Evaluating the carbon footprint of code implementation |
description |
This project undertakes the evaluation of the carbon footprint of large language models (LLMs). Three models are evaluated – Meta’s LLaMA-2 (7-billion parameter configuration), Mistral (7-billion parameter configuration), and Google’s Gemma (2-billion and 7-billion parameter configurations). The emissions generated by these models in the fine-tuning phase are evaluated for three tasks – question answering, text summarisation, and sentiment analysis. This report also examines differences in emissions generated across GPUs. The project ultimately explores the impact of model optimisation on emissions and provides corresponding recommendations to reduce the carbon footprint of the selected models. |
author2 |
Lim Wei Yang Bryan |
author_facet |
Lim Wei Yang Bryan Tar, Sreeja |
format |
Final Year Project |
author |
Tar, Sreeja |
author_sort |
Tar, Sreeja |
title |
Evaluating the carbon footprint of code implementation |
title_short |
Evaluating the carbon footprint of code implementation |
title_full |
Evaluating the carbon footprint of code implementation |
title_fullStr |
Evaluating the carbon footprint of code implementation |
title_full_unstemmed |
Evaluating the carbon footprint of code implementation |
title_sort |
evaluating the carbon footprint of code implementation |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/181172 |
_version_ |
1816858966047588352 |