THE EFFECT OF PREPROCESSING TECHNIQUES ON STACKED GENERALIZATION AND STAND-ALONE METHOD FOR E-COMMERCE DEMAND PREDICTION

The competition between sellers of e-commerce in Indonesia are getting more intense. One of the ways for the sellers to stay competing in the industry is by managing their product inventory through demand prediction. A lot of research related to demand prediction had been conducted. On those rese...

Full description

Saved in:
Bibliographic Details
Main Author: Athena M. S., Cynthia
Format: Final Project
Language:Indonesia
Online Access:https://digilib.itb.ac.id/gdl/view/66588
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Institut Teknologi Bandung
Language: Indonesia
Description
Summary:The competition between sellers of e-commerce in Indonesia are getting more intense. One of the ways for the sellers to stay competing in the industry is by managing their product inventory through demand prediction. A lot of research related to demand prediction had been conducted. On those research, preprocessing techniques were always implemented, but there hasn’t been a statement or a main guide related to the effect of preprocessing techniques on ecommerce demand prediction, like the research that was conducted on 2022 for steel industry. In this paper, multiple scenarios will be conducted to see the effects of preprocessing techniques on stacked generalization and stand-alone method using MLP (Multi Layer Perceptron) and XGBoost (Xtreme Gradient Boosting). Stacked generalization that is used is the one that was developed by Irem islek and Sule Gunduz on 2017. The preprocessing techniques that are implemented including data scaling, and feature interaction constraint on XGBoost. Hyperparameter tuning is also implemented. It is found that data scaling can increase the performance of stacked generalization with MLP on level 1. The training time will also be faster if there is a significant improvement from highly skewed to normal distribution data. With feature interaction constraint, the training time of stacked generalization and stand-alone method can become faster with the same or better prediction performance, if the constraint is in accordance with the problem. Stacked generalization, especially with XGBoost on level 1, can result in a better prediction than stand-alone MLP with hyperparameter tuning and data scaling. If the data doesn’t have a high skewness, stacked generalization method can minimalize the error made by unsuitable parameter in stand-alone XGBoost.