Skip main navigation

Performance of the multilayer perceptron

Ian Witten reviews the performance of multilayer perceptrons in the preceding experiments

Let’s review the performance of multilayer perceptrons in the preceding quiz.

First, 2 hidden layers never significantly outperform 1 hidden layer.

Ignoring significance:

  • 2 hidden layers are best for 1 dataset

  • 1 hidden layer is best for 2 datasets

  • 0 hidden layers are best for 3 datasets.

It’s interesting to look at the time taken for training, UserCPU_Time_training. Overall, 2 hidden layers take up to twice as long as 1 hidden layer, which takes between 2 and 12 times as long as 0 hidden layers.

On 4 of the 6 datasets, all three versions of MultilayerPerceptron (0, 1, and 2 hidden layers) are outperformed by much faster methods:

  • roundly outperformed by J48 on breast-cancer (74% correct vs 67% correct for the default configuration of MultilayerPerceptron)

  • outperformed by SMO on credit-g and diabetes (75% vs 72%, and 77% vs 75% respectively)

  • outperformed by IBk on glass (70% vs 67%).

MultilayerPerceptron (default configuration with 1 hidden layer) has two marginal successes:

  • on iris, it’s a shade better than its nearest competitor, SMO (97% vs 96%)

  • on ionosphere, it’s a shade better than its nearest competitor, AdaBoostM1 (91.1% vs 90.9%).

This article is from the free online

More Data Mining with Weka

Created by
FutureLearn - Learning For Life

Reach your personal and professional goals

Unlock access to hundreds of expert online courses and degrees from top universities and educators to gain accredited qualifications and professional CV-building certificates.

Join over 18 million learners to launch, switch or build upon your career, all at your own pace, across a wide range of topic areas.

Start Learning now