Evaluating the carbon footprint of code implementation

This project undertakes the evaluation of the carbon footprint of large language models (LLMs). Three models are evaluated – Meta’s LLaMA-2 (7-billion parameter configuration), Mistral (7-billion parameter configuration), and Google’s Gemma (2-billion and 7-billion parameter configurations). The emi...

Full description

Saved in:
Bibliographic Details
Main Author: Tar, Sreeja
Other Authors: Lim Wei Yang Bryan
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/181172
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-181172
record_format dspace
spelling sg-ntu-dr.10356-1811722024-11-18T01:35:30Z Evaluating the carbon footprint of code implementation Tar, Sreeja Lim Wei Yang Bryan College of Computing and Data Science bryan.limwy@ntu.edu.sg Computer and Information Science Large Language Models (LLMs) Carbon footprint Fine-tuning Natural Language Processing (NLP) Emissions Energy consumption This project undertakes the evaluation of the carbon footprint of large language models (LLMs). Three models are evaluated – Meta’s LLaMA-2 (7-billion parameter configuration), Mistral (7-billion parameter configuration), and Google’s Gemma (2-billion and 7-billion parameter configurations). The emissions generated by these models in the fine-tuning phase are evaluated for three tasks – question answering, text summarisation, and sentiment analysis. This report also examines differences in emissions generated across GPUs. The project ultimately explores the impact of model optimisation on emissions and provides corresponding recommendations to reduce the carbon footprint of the selected models. Bachelor's degree 2024-11-18T01:35:06Z 2024-11-18T01:35:06Z 2024 Final Year Project (FYP) Tar, S. (2024). Evaluating the carbon footprint of code implementation. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181172 https://hdl.handle.net/10356/181172 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
Large Language Models (LLMs)
Carbon footprint
Fine-tuning
Natural Language Processing (NLP)
Emissions
Energy consumption
spellingShingle Computer and Information Science
Large Language Models (LLMs)
Carbon footprint
Fine-tuning
Natural Language Processing (NLP)
Emissions
Energy consumption
Tar, Sreeja
Evaluating the carbon footprint of code implementation
description This project undertakes the evaluation of the carbon footprint of large language models (LLMs). Three models are evaluated – Meta’s LLaMA-2 (7-billion parameter configuration), Mistral (7-billion parameter configuration), and Google’s Gemma (2-billion and 7-billion parameter configurations). The emissions generated by these models in the fine-tuning phase are evaluated for three tasks – question answering, text summarisation, and sentiment analysis. This report also examines differences in emissions generated across GPUs. The project ultimately explores the impact of model optimisation on emissions and provides corresponding recommendations to reduce the carbon footprint of the selected models.
author2 Lim Wei Yang Bryan
author_facet Lim Wei Yang Bryan
Tar, Sreeja
format Final Year Project
author Tar, Sreeja
author_sort Tar, Sreeja
title Evaluating the carbon footprint of code implementation
title_short Evaluating the carbon footprint of code implementation
title_full Evaluating the carbon footprint of code implementation
title_fullStr Evaluating the carbon footprint of code implementation
title_full_unstemmed Evaluating the carbon footprint of code implementation
title_sort evaluating the carbon footprint of code implementation
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/181172
_version_ 1816858966047588352