Page 218 - 2024-Vol20-Issue2
P. 218

214 |                                                                                     Gochhait, Sharma & Bachute

                  TABLE VI.                                                      TABLE VII.
TRAINING OPTIONS FOR ODISHA STATE                                TRAINING LAYERS FOR DELHI STATE

              Layer               Options                                   Layer              Parameters/Options
              adam                     -                            SequenceInputLayer    numFeatures, ”Name”, ”input”
      GradientThreshold               1                             Convolution1dLayer
        InitialLearnRate                                         BatchNormalizationLayer     11, 96, ’Padding’, ’same’
          MaxEpochs                 0.001                                                                   -
       SequenceLength               1000                                 ReLULayer                          -
             Epsilon             ”longest”                          Convolution1dLayer
       L2Regularization           1 × 10-8                       BatchNormalizationLayer    20, 180, ’Padding’, ’same’
             Shuffle               0.0001                                                                   -
    GradientDecayFactor       ”every-epoch”                              ReLULayer                          -
SquaredGradientDecayFactor           0.9                            Convolution1dLayer
    LearnRateDropFactor             0.999                        BatchNormalizationLayer    30, 300, ’Padding’, ’same’
    LearnRateDropPeriod              0.1                                                                    -
 GradientThresholdMethod                                                 ReLULayer                          -
  ResetInputNormalization             10                            Convolution1dLayer
              Plots              ”l2norm”                        BatchNormalizationLayer    32, 320, ’Padding’, ’same’
                                                                                                            -
                                     true                                ReLULayer                          -
                            ”training-progress”                         DropoutLayer                       0.2
                                                                       BiLSTMLayer
dataset, with an 80% training and 20% validation ratio from             DropoutLayer      100, ’OutputMode’, ’sequence’
the available datasets.                                                BiLSTMLayer                         0.1
                                                                        DropoutLayer
       III. DEEP LEARNING MODEL FOR                                    BiLSTMLayer        105, ’OutputMode’, ’sequence’
              LONG-TERM ELECTRICITY                                     DropoutLayer                       0.2
                       FORECASTING                                 FullyConnectedLayer
                                                                      RegressionLayer     110, ’OutputMode’, ’sequence’
Recent advancements in deep learning algorithms, such as                                                   0.2
convolutional neural net- works (CNNs) and recurrent neural                                                 1
networks (RNNs), have shown remarkable efficacy in various                                                  -
domains. Among these, the Bidirectional Long Short-Term
Memory (Bi- LSTM) model has gained significant attention                         TABLE VIII.
as an effective RNN structure for time series prediction. Ad-    TRAINING OPTIONS FOR DELHI STATE
ditionally, there is a growing trend of leveraging 1D CNNs
orcombining both 1D CNN and Bi-LSTM algorithms to en-                        Options                Values
hance forecasting accuracy. Comparative studies have consis-           GradientThreshold               1
tently demonstrated the superior performance of these models
compared to conventional statistical or machine-learning mod-            InitialLearnRate           0.001
els.                                                                       MaxEpochs                 1000
In this section, we present a comprehensive overview of the                                       ’longest’
theoretical foundations underlying these neural networks. We            SequenceLength            1 × 10-8
delve into the mechanisms and architectural details of CNNs                   Epsilon               0.0001
and RNNs, with a particular focus on the Bi-LSTM model.                                        ’every-epoch’
The aim is to provide a clear understanding of the proposed             L2Regularization              0.9
models and their potential benefits for long-term electricity                 Shuffle               0.999
forecasting in the context of our case study of Orissa andDelhi                                       0.1
States.                                                              GradientDecayFactor              10
                                                                 SquaredGradientDecayFactor       ’l2norm’
A. CNN-BI LSTM hybrid model with deep 1D modeling                                                    true
Multivariate 1D time-series signals can be accurately pre-           LearnRateDropFactor     ’training-progress’
dicted by combining a hybrid model of 1D-CNN and BiL-                LearnRateDropPeriod
                                                                  GradientThresholdMethod
                                                                   ResetInputNormalization

                                                                               Plots

                                                                 STM. This approach has been extensively studied and shown
                                                                 promising results in various domains such as weather predic-
                                                                 tion, speech recognition, stock price forecasting, and power
                                                                 usage prediction. The combinationof Convolutional Neural
                                                                 Networks (CNNs) and Long Short-Term Memory (LSTM)
   213   214   215   216   217   218   219   220   221   222   223