From 76332c00661f39493651bfce2d1493a2a86e786e Mon Sep 17 00:00:00 2001 From: KOSASIH Date: Thu, 25 Jul 2024 15:31:39 +0700 Subject: [PATCH] Create NLP.java --- .../src/main/java/com/sidra/NLP.java | 41 +++++++++++++++++++ 1 file changed, 41 insertions(+) create mode 100644 sidra_chain_integration/project/sidra_nexus/src/main/java/com/sidra/NLP.java diff --git a/sidra_chain_integration/project/sidra_nexus/src/main/java/com/sidra/NLP.java b/sidra_chain_integration/project/sidra_nexus/src/main/java/com/sidra/NLP.java new file mode 100644 index 000000000..01ee7c26f --- /dev/null +++ b/sidra_chain_integration/project/sidra_nexus/src/main/java/com/sidra/NLP.java @@ -0,0 +1,41 @@ +package com.sidra; + +import org.deeplearning4j.nn.conf.NeuralNetConfiguration; +import org.deeplearning4j.nn.conf.layers.LSTM; +import org.deeplearning4j.nn.conf.layers.RnnOutputLayer; +import org.deeplearning4j.nn.multilayer.MultiLayerNetwork; +import org.deeplearning4j.nn.weights.WeightInit; +import org.nd4j.linalg.activations.Activation; +import org.nd4j.linalg.dataset.api.iterator.DataSetIterator; +import org.nd4j.linalg.lossfunctions.LossFunctions; + +public class NLP { + public static void main(String[] args) { + // Set up a neural network for NLP + NeuralNetConfiguration config = new NeuralNetConfiguration.Builder() + .seed(42) + .weightInit(WeightInit.XAVIER) + .updater(new Nesterovs(0.01)) + .list() + .layer(new LSTM.Builder() + .nIn(100) + .nOut(100) + .activation(Activation.TANH) + .build()) + .layer(new RnnOutputLayer.Builder(LossFunctions.LossFunction.MCXENT) + .activation(Activation.SOFTMAX) + .nIn(100) + .nOut(10) + .build()) + .pretrain(false) + .backprop(true) + .build(); + + MultiLayerNetwork model = new MultiLayerNetwork(config); + model.init(); + + // Train the model + DataSetIterator iterator = new DataSetIterator(); + model.fit(iterator); + } +}