You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
mindspore/model_zoo/official/cv/alexnet
caojiewen cad462902a
fixed the code spell errors.
4 years ago
..
scripts gpu fix padding and dataset_heleper 4 years ago
src change md numa link to dllload 4 years ago
README.md 1. fixed for markdownlint errors. 4 years ago
README_CN.md 1. fixed for markdownlint errors. 4 years ago
eval.py fixed the code spell errors. 4 years ago
export.py fix GPU device_id bug 4 years ago
mindspore_hub_conf.py add hub config for lenet and alexnet 4 years ago
requirements.txt add script shell 5 years ago
train.py fixed the code spell errors. 4 years ago

README.md

Contents

AlexNet Description

AlexNet was proposed in 2012, one of the most influential neural networks. It got big success in ImageNet Dataset recognition than other models.

Paper: Krizhevsky A, Sutskever I, Hinton G E. ImageNet Classification with Deep ConvolutionalNeural Networks. Advances In Neural Information Processing Systems. 2012.

Model Architecture

AlexNet composition consists of 5 convolutional layers and 3 fully connected layers. Multiple convolutional kernels can extract interesting features in images and get more accurate classification.

Dataset

Note that you can run the scripts based on the dataset mentioned in original paper or widely used in relevant domain/network architecture. In the following sections, we will introduce how to run the scripts using the related dataset below.

Dataset used: CIFAR-10

  • Dataset size175M60,000 32*32 colorful images in 10 classes
    • Train146M50,000 images
    • Test29.3M10,000 images
  • Data formatbinary files
    • NoteData will be processed in dataset.py
  • Download the dataset, the directory structure is as follows:
├─cifar-10-batches-bin
│
└─cifar-10-verify-bin

Environment Requirements

Quick Start

After installing MindSpore via the official website, you can start training and evaluation as follows:

# enter script dir, train AlexNet
sh run_standalone_train_ascend.sh [DATA_PATH] [CKPT_SAVE_PATH]
# enter script dir, evaluate AlexNet
sh run_standalone_eval_ascend.sh [DATA_PATH] [CKPT_NAME]

Script Description

Script and Sample Code

├── cv
    ├── alexnet
        ├── README.md                    // descriptions about alexnet
        ├── requirements.txt             // package needed
        ├── scripts
        │   ├──run_standalone_train_gpu.sh             // train in gpu
        │   ├──run_standalone_train_ascend.sh          // train in ascend
        │   ├──run_standalone_eval_gpu.sh             //  evaluate in gpu
        │   ├──run_standalone_eval_ascend.sh          //  evaluate in ascend
        ├── src
        │   ├──dataset.py             // creating dataset
        │   ├──alexnet.py              // alexnet architecture
        │   ├──config.py            // parameter configuration
        ├── train.py               // training script
        ├── eval.py               //  evaluation script

Script Parameters

Major parameters in train.py and config.py as follows:

--data_path: The absolute full path to the train and evaluation datasets.
--epoch_size: Total training epochs.
--batch_size: Training batch size.
--image_height: Image height used as input to the model.
--image_width: Image width used as input the model.
--device_target: Device where the code will be implemented. Optional values are "Ascend", "GPU".
--checkpoint_path: The absolute full path to the checkpoint file saved after training.
--data_path: Path where the dataset is saved

Training Process

Training

  • running on Ascend

    python train.py --data_path cifar-10-batches-bin --ckpt_path ckpt > log 2>&1 &
    # or enter script dir, and run the script
    sh run_standalone_train_ascend.sh cifar-10-batches-bin ckpt
    

    After training, the loss value will be achieved as follows:

    # grep "loss is " log
    epoch: 1 step: 1, loss is 2.2791853
    ...
    epoch: 1 step: 1536, loss is 1.9366643
    epoch: 1 step: 1537, loss is 1.6983616
    epoch: 1 step: 1538, loss is 1.0221305
    ...
    

    The model checkpoint will be saved in the current directory.

  • running on GPU

    python train.py --device_target "GPU" --data_path cifar-10-batches-bin --ckpt_path ckpt > log 2>&1 &
    # or enter script dir, and run the script
    sh run_standalone_train_for_gpu.sh cifar-10-batches-bin ckpt
    

    After training, the loss value will be achieved as follows:

    # grep "loss is " log
    epoch: 1 step: 1, loss is 2.3125906
    ...
    epoch: 30 step: 1560, loss is 0.6687547
    epoch: 30 step: 1561, loss is 0.20055409
    epoch: 30 step: 1561, loss is 0.103845775
    

Evaluation Process

Evaluation

Before running the command below, please check the checkpoint path used for evaluation.

  • running on Ascend

    python eval.py --data_path cifar-10-verify-bin --ckpt_path ckpt/checkpoint_alexnet-1_1562.ckpt > eval_log.txt 2>&1 &
    # or enter script dir, and run the script
    sh run_standalone_eval_ascend.sh cifar-10-verify-bin ckpt/checkpoint_alexnet-1_1562.ckpt
    

    You can view the results through the file "eval_log". The accuracy of the test dataset will be as follows:

    # grep "Accuracy: " eval_log
    'Accuracy': 0.8832
    
  • running on GPU

    python eval.py --device_target "GPU" --data_path cifar-10-verify-bin --ckpt_path ckpt/checkpoint_alexnet-30_1562.ckpt > eval_log 2>&1 &
    # or enter script dir, and run the script
    sh run_standalone_eval_for_gpu.sh cifar-10-verify-bin ckpt/checkpoint_alexnet-30_1562.ckpt
    

    You can view the results through the file "eval_log". The accuracy of the test dataset will be as follows:

    # grep "Accuracy: " eval_log
    'Accuracy': 0.88512
    

Model Description

Performance

Evaluation Performance

Parameters Ascend GPU
Resource Ascend 910; CPU 2.60GHz, 192cores; Memory, 755G NV SMX2 V100-32G
uploaded Date 06/09/2020 (month/day/year) 17/09/2020 (month/day/year)
MindSpore Version 1.0.0 0.7.0-beta
Dataset CIFAR-10 CIFAR-10
Training Parameters epoch=30, steps=1562, batch_size = 32, lr=0.002 epoch=30, steps=1562, batch_size = 32, lr=0.002
Optimizer Momentum Momentum
Loss Function Softmax Cross Entropy Softmax Cross Entropy
outputs probability probability
Loss 0.08 0.01
Speed 7.3 ms/step 16.8 ms/step
Total time 6 mins 14 mins
Checkpoint for Fine tuning 445M (.ckpt file) 445M (.ckpt file)
Scripts AlexNet Script AlexNet Script

Description of Random Situation

In dataset.py, we set the seed inside create_dataset function.

ModelZoo Homepage

Please check the official homepage.