Task1: Training masked autoencoder on PASCAL VOC 2007 dataset Dataset preparation: Download the PASCAL VOC 2007 dataset from here. (http://host.robots.ox.ac.uk/pascal/VOC/voc2007/) Preprocess the dataset by resizing the images to a fixed size and normalizing the pixel values. Dataset split: (Run Task.1.3(a) with both the following splits and choose the best one for Task.1.3(b), Task.1.3(c) and next experiments.) Use 80-10-10, train-val-test split Use 70-10-20, train-val-test split Architecture: An autoencoder with three hidden layers with the following bottleneck dimension (on denoising task, add gaussian noise to the input image, and the output should be denoised image) 256 (run using both the splits in 2, and choose the best one for the further set of experiments) 128 64 32 16 Choose the best bottleneck dimension, and re-run the autoencoder using masking strategy: (mask the following % of pixels in the image, i.e., set the pixel value to (0,0,0)) 20% 40% 60% 80%. Plot reconstruction error for every autoencoder model. Evaluation: Report MSE (mean square error), MAE (mean absolute error) for all models. Comment on which metric is more useful in judging the quality . Visualize and compare the original images, masked images, and reconstructed images. Use any other metric of your choice (apart from MSE, MAE) to judge the image quality. Use the best split, and best masking strategy (with best encoding dimension) for Task 2. Task2: Fine-tuning a pre-trained autoencoder on STL-10 dataset Dataset: Download the STL-10 dataset from here (https://cs.stanford.edu/~acoates/stl10/)and pre-process as required. Use the encoder of the above-pretrained autoencoder (Refer to task 1.9) as a feature extractor. Build a downstream task classifier (a MLP) with (100% training samples) and choose the best one among a, and b for Task 2.4. # hidden_layer = 3 # hidden_layer = 5 Use the best from Task 2.3.a and 2.3.b, fine-tune the classifier on the STL-10 dataset with the following % of training samples 1% 10% 20% 40%. 60 % Prepare confusion matrix and AUC-ROC curve to evaluate the model performance (for Task 2.3.(a-b) and Task 2.4.(a-e)). Implement a different architecture than the architecture described in Task.2.3.(a-b), that improves the result. Task3: Report hyperparameters clearly for every model No. of epochs Learning rate Optimizer Loss function. Dimensions Report and justify your choice of best (for split, for bottleneck dimension, masking strategy etc.) Observations about results, any additional analysis over results.
Task1: Training masked autoencoder on PASCAL VOC 2007 dataset Dataset preparation: Download the PASCAL VOC 2007 dataset from here. (http://host.robots.ox.ac.uk/pascal/VOC/voc2007/) Preprocess the dataset by resizing the images to a fixed size and normalizing the pixel values. Dataset split: (Run Task.1.3(a) with both the following splits and choose the best one for Task.1.3(b), Task.1.3(c) and next experiments.) Use 80-10-10, train-val-test split Use 70-10-20, train-val-test split Architecture: An autoencoder with three hidden layers with the following bottleneck dimension (on denoising task, add gaussian noise to the input image, and the output should be denoised image) 256 (run using both the splits in 2, and choose the best one for the further set of experiments) 128 64 32 16 Choose the best bottleneck dimension, and re-run the autoencoder using masking strategy: (mask the following % of pixels in the image, i.e., set the pixel value to (0,0,0)) 20% 40% 60% 80%. Plot reconstruction error for every autoencoder model. Evaluation: Report MSE (mean square error), MAE (mean absolute error) for all models. Comment on which metric is more useful in judging the quality . Visualize and compare the original images, masked images, and reconstructed images. Use any other metric of your choice (apart from MSE, MAE) to judge the image quality. Use the best split, and best masking strategy (with best encoding dimension) for Task 2. Task2: Fine-tuning a pre-trained autoencoder on STL-10 dataset Dataset: Download the STL-10 dataset from here (https://cs.stanford.edu/~acoates/stl10/)and pre-process as required. Use the encoder of the above-pretrained autoencoder (Refer to task 1.9) as a feature extractor. Build a downstream task classifier (a MLP) with (100% training samples) and choose the best one among a, and b for Task 2.4. # hidden_layer = 3 # hidden_layer = 5 Use the best from Task 2.3.a and 2.3.b, fine-tune the classifier on the STL-10 dataset with the following % of training samples 1% 10% 20% 40%. 60 % Prepare confusion matrix and AUC-ROC curve to evaluate the model performance (for Task 2.3.(a-b) and Task 2.4.(a-e)). Implement a different architecture than the architecture described in Task.2.3.(a-b), that improves the result. Task3: Report hyperparameters clearly for every model No. of epochs Learning rate Optimizer Loss function. Dimensions Report and justify your choice of best (for split, for bottleneck dimension, masking strategy etc.) Observations about results, any additional analysis over results.
Database System Concepts
7th Edition
ISBN:9780078022159
Author:Abraham Silberschatz Professor, Henry F. Korth, S. Sudarshan
Publisher:Abraham Silberschatz Professor, Henry F. Korth, S. Sudarshan
Chapter1: Introduction
Section: Chapter Questions
Problem 1PE
Related questions
Question
Task1: Training masked autoencoder on PASCAL VOC 2007 dataset
- Dataset preparation:
- Download the PASCAL VOC 2007 dataset from here. (http://host.robots.ox.ac.uk/pascal/VOC/voc2007/)
- Preprocess the dataset by resizing the images to a fixed size and normalizing the pixel values.
- Dataset split: (Run Task.1.3(a) with both the following splits and choose the best one for Task.1.3(b), Task.1.3(c) and next experiments.)
- Use 80-10-10, train-val-test split
- Use 70-10-20, train-val-test split
- Architecture: An autoencoder with three hidden layers with the following bottleneck dimension (on denoising task, add gaussian noise to the input image, and the output should be denoised image)
- 256 (run using both the splits in 2, and choose the best one for the further set of experiments)
- 128
- 64
- 32
- 16
- Choose the best bottleneck dimension, and re-run the autoencoder using masking strategy: (mask the following % of pixels in the image, i.e., set the pixel value to (0,0,0))
- 20%
- 40%
- 60%
- 80%.
- Plot reconstruction error for every autoencoder model.
- Evaluation: Report MSE (mean square error), MAE (mean absolute error) for all models. Comment on which metric is more useful in judging the quality .
- Visualize and compare the original images, masked images, and reconstructed images.
- Use any other metric of your choice (apart from MSE, MAE) to judge the image quality.
- Use the best split, and best masking strategy (with best encoding dimension) for Task 2.
Task2: Fine-tuning a pre-trained autoencoder on STL-10 dataset
- Dataset: Download the STL-10 dataset from here (https://cs.stanford.edu/~acoates/stl10/)and pre-process as required.
- Use the encoder of the above-pretrained autoencoder (Refer to task 1.9) as a feature extractor.
- Build a downstream task classifier (a MLP) with (100% training samples) and choose the best one among a, and b for Task 2.4.
- # hidden_layer = 3
- # hidden_layer = 5
- Use the best from Task 2.3.a and 2.3.b, fine-tune the classifier on the STL-10 dataset with the following % of training samples
- 1%
- 10%
- 20%
- 40%.
- 60 %
- Prepare confusion matrix and AUC-ROC curve to evaluate the model performance (for Task 2.3.(a-b) and Task 2.4.(a-e)).
- Implement a different architecture than the architecture described in Task.2.3.(a-b), that improves the result.
Task3:
- Report hyperparameters clearly for every model
- No. of epochs
- Learning rate
- Optimizer
- Loss function.
- Dimensions
- Report and justify your choice of best (for split, for bottleneck dimension, masking strategy etc.)
- Observations about results, any additional analysis over results.
Expert Solution
This question has been solved!
Explore an expertly crafted, step-by-step solution for a thorough understanding of key concepts.
This is a popular solution!
Trending now
This is a popular solution!
Step by step
Solved in 3 steps
Knowledge Booster
Learn more about
Need a deep-dive on the concept behind this application? Look no further. Learn more about this topic, computer-science and related others by exploring similar questions and additional content below.Recommended textbooks for you
Database System Concepts
Computer Science
ISBN:
9780078022159
Author:
Abraham Silberschatz Professor, Henry F. Korth, S. Sudarshan
Publisher:
McGraw-Hill Education
Starting Out with Python (4th Edition)
Computer Science
ISBN:
9780134444321
Author:
Tony Gaddis
Publisher:
PEARSON
Digital Fundamentals (11th Edition)
Computer Science
ISBN:
9780132737968
Author:
Thomas L. Floyd
Publisher:
PEARSON
Database System Concepts
Computer Science
ISBN:
9780078022159
Author:
Abraham Silberschatz Professor, Henry F. Korth, S. Sudarshan
Publisher:
McGraw-Hill Education
Starting Out with Python (4th Edition)
Computer Science
ISBN:
9780134444321
Author:
Tony Gaddis
Publisher:
PEARSON
Digital Fundamentals (11th Edition)
Computer Science
ISBN:
9780132737968
Author:
Thomas L. Floyd
Publisher:
PEARSON
C How to Program (8th Edition)
Computer Science
ISBN:
9780133976892
Author:
Paul J. Deitel, Harvey Deitel
Publisher:
PEARSON
Database Systems: Design, Implementation, & Manag…
Computer Science
ISBN:
9781337627900
Author:
Carlos Coronel, Steven Morris
Publisher:
Cengage Learning
Programmable Logic Controllers
Computer Science
ISBN:
9780073373843
Author:
Frank D. Petruzella
Publisher:
McGraw-Hill Education