Skip to main content

Table 3 Parameter settings of classifiers

From: Identifying tweets of personal health experience through word embedding and LSTM neural network

Classifier

Parameter Settings

Logistic Regression

penalty:'l2ā€™, tolā€‰=ā€‰1e-4, Cā€‰=ā€‰1.0, solver:'liblinearā€™,max_iterā€‰=ā€‰100

Decision Tree (J48)

criterionā€‰=ā€‰ā€˜entropyā€™, max_depthā€‰=ā€‰30, min_samples_splitā€‰=ā€‰2, min_samples_leafā€‰=ā€‰1

KNN

n_neighborsā€‰=ā€‰1, pā€‰=ā€‰2, metricā€‰=ā€‰ā€˜minkowskiā€™,algorithmā€‰=ā€‰ā€˜autoā€™

SVM

Cā€‰=ā€‰1.0, kernelā€‰=ā€‰ā€˜rbfā€™, tolā€‰=ā€‰1e-4, gommaā€‰=ā€‰0.001

BoW + Logistic Regr.

Cā€‰=ā€‰1000, random_stateā€‰=ā€‰0

Word Embedding + LSTM

In LSTM layer, the input and output dimensions: 128, L2 for regularizer, and the parameter for L2: 0.01. 30% of training dataset was used as validation dataset. Class weight for PET class: 6547/2650, and for non-PET class: 2650/6547