From 959862e4231562b32036078b9ec99dd7a7769163 Mon Sep 17 00:00:00 2001 From: Jean-Luc Parouty <Jean-Luc.Parouty@grenoble-inp.fr> Date: Fri, 8 Jan 2021 22:54:57 +0100 Subject: [PATCH] Update notebooks descriptions --- GTSRB/01-Preparation-of-data.ipynb | 4 +- GTSRB/02-First-convolutions.ipynb | 4 +- GTSRB/03-Tracking-and-visualizing.ipynb | 4 +- GTSRB/04-Data-augmentation.ipynb | 4 +- GTSRB/05-Full-convolutions.ipynb | 2 +- GTSRB/06-Notebook-as-a-batch.ipynb | 2 +- GTSRB/07-Show-report.ipynb | 4 +- GTSRB/batch_oar.sh | 4 +- GTSRB/batch_slurm.sh | 2 +- IMDB/01-Embedding-Keras.ipynb | 6 +-- IMDB/02-Prediction.ipynb | 4 +- IMDB/03-LSTM-Keras.ipynb | 4 +- README.ipynb | 53 ++++++++++++++----------- README.md | 40 +++++++++---------- fidle/01 - Set and reset.ipynb | 44 ++++++++++---------- fidle/log/catalog.json | 40 +++++++++---------- 16 files changed, 115 insertions(+), 106 deletions(-) diff --git a/GTSRB/01-Preparation-of-data.ipynb b/GTSRB/01-Preparation-of-data.ipynb index b2ce2a7..ddad62e 100644 --- a/GTSRB/01-Preparation-of-data.ipynb +++ b/GTSRB/01-Preparation-of-data.ipynb @@ -6,8 +6,8 @@ "source": [ "<img width=\"800px\" src=\"../fidle/img/00-Fidle-header-01.svg\"></img>\n", "\n", - "# <!-- TITLE --> [GTSRB1] - CNN with GTSRB dataset - Data analysis and preparation\n", - "<!-- DESC --> Episode 1 : Data analysis and creation of a usable dataset\n", + "# <!-- TITLE --> [GTSRB1] - Dataset analysis and preparation\n", + "<!-- DESC --> Episode 1 : Analysis of the GTSRB dataset and creation of an enhanced dataset\n", "<!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) -->\n", "\n", "## Objectives :\n", diff --git a/GTSRB/02-First-convolutions.ipynb b/GTSRB/02-First-convolutions.ipynb index 4d478af..37e259b 100644 --- a/GTSRB/02-First-convolutions.ipynb +++ b/GTSRB/02-First-convolutions.ipynb @@ -6,8 +6,8 @@ "source": [ "<img width=\"800px\" src=\"../fidle/img/00-Fidle-header-01.svg\"></img>\n", "\n", - "# <!-- TITLE --> [GTSRB2] - CNN with GTSRB dataset - First convolutions\n", - "<!-- DESC --> Episode 2 : First convolutions and first results\n", + "# <!-- TITLE --> [GTSRB2] - First convolutions\n", + "<!-- DESC --> Episode 2 : First convolutions and first classification of our traffic signs\n", "<!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) -->\n", "\n", "## Objectives :\n", diff --git a/GTSRB/03-Tracking-and-visualizing.ipynb b/GTSRB/03-Tracking-and-visualizing.ipynb index 1ada67a..8d8f089 100644 --- a/GTSRB/03-Tracking-and-visualizing.ipynb +++ b/GTSRB/03-Tracking-and-visualizing.ipynb @@ -6,8 +6,8 @@ "source": [ "<img width=\"800px\" src=\"../fidle/img/00-Fidle-header-01.svg\"></img>\n", "\n", - "# <!-- TITLE --> [GTSRB3] - CNN with GTSRB dataset - Monitoring \n", - "<!-- DESC --> Episode 3 : Monitoring and analysing training, managing checkpoints\n", + "# <!-- TITLE --> [GTSRB3] - Training monitoring\n", + "<!-- DESC --> Episode 3 : Monitoring, analysis and check points during a training session\n", "<!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) -->\n", "\n", "## Objectives :\n", diff --git a/GTSRB/04-Data-augmentation.ipynb b/GTSRB/04-Data-augmentation.ipynb index 5fea43a..fd336e5 100644 --- a/GTSRB/04-Data-augmentation.ipynb +++ b/GTSRB/04-Data-augmentation.ipynb @@ -6,8 +6,8 @@ "source": [ "<img width=\"800px\" src=\"../fidle/img/00-Fidle-header-01.svg\"></img>\n", "\n", - "# <!-- TITLE --> [GTSRB4] - CNN with GTSRB dataset - Data augmentation \n", - "<!-- DESC --> Episode 4 : Improving the results with data augmentation\n", + "# <!-- TITLE --> [GTSRB4] - Data augmentation \n", + "<!-- DESC --> Episode 4 : Adding data by data augmentation when we lack it, to improve our results\n", "<!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) -->\n", "\n", "## Objectives :\n", diff --git a/GTSRB/05-Full-convolutions.ipynb b/GTSRB/05-Full-convolutions.ipynb index 2b7274f..1c169c9 100644 --- a/GTSRB/05-Full-convolutions.ipynb +++ b/GTSRB/05-Full-convolutions.ipynb @@ -6,7 +6,7 @@ "source": [ "<img width=\"800px\" src=\"../fidle/img/00-Fidle-header-01.svg\"></img>\n", "\n", - "# <!-- TITLE --> [GTSRB5] - CNN with GTSRB dataset - Full convolutions \n", + "# <!-- TITLE --> [GTSRB5] - Full convolutions\n", "<!-- DESC --> Episode 5 : A lot of models, a lot of datasets and a lot of results.\n", "<!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) -->\n", "\n", diff --git a/GTSRB/06-Notebook-as-a-batch.ipynb b/GTSRB/06-Notebook-as-a-batch.ipynb index 9ef2ef3..8902e0e 100644 --- a/GTSRB/06-Notebook-as-a-batch.ipynb +++ b/GTSRB/06-Notebook-as-a-batch.ipynb @@ -7,7 +7,7 @@ "<img width=\"800px\" src=\"../fidle/img/00-Fidle-header-01.svg\"></img>\n", "\n", "# <!-- TITLE --> [GTSRB6] - Full convolutions as a batch\n", - "<!-- DESC --> Episode 6 : Run Full convolution notebook as a batch\n", + "<!-- DESC --> Episode 6 : To compute bigger, use your notebook in batch mode\n", "<!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) -->\n", "\n", "## Objectives :\n", diff --git a/GTSRB/07-Show-report.ipynb b/GTSRB/07-Show-report.ipynb index 0b1b347..0999929 100644 --- a/GTSRB/07-Show-report.ipynb +++ b/GTSRB/07-Show-report.ipynb @@ -6,8 +6,8 @@ "source": [ "<img width=\"800px\" src=\"../fidle/img/00-Fidle-header-01.svg\"></img>\n", "\n", - "# <!-- TITLE --> [GTSRB7] - CNN with GTSRB dataset - Show reports\n", - "<!-- DESC --> Episode 7 : Displaying a jobs report\n", + "# <!-- TITLE --> [GTSRB7] - Batch reportss\n", + "<!-- DESC --> Episode 7 : Displaying our jobs report, and the winner is...\n", "<!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) -->\n", "\n", "## Objectives :\n", diff --git a/GTSRB/batch_oar.sh b/GTSRB/batch_oar.sh index 272acfe..0819741 100755 --- a/GTSRB/batch_oar.sh +++ b/GTSRB/batch_oar.sh @@ -19,8 +19,8 @@ # Fidle at GRICAD # ----------------------------------------------- # -# <!-- TITLE --> [GTSRB10] - OAR batch submission -# <!-- DESC --> Bash script for OAR batch submission of GTSRB notebook +# <!-- TITLE --> [GTSRB10] - OAR batch script submission +# <!-- DESC --> Bash script for an OAR batch submission of an ipython code # <!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) --> # ==== Notebook parameters ========================================= diff --git a/GTSRB/batch_slurm.sh b/GTSRB/batch_slurm.sh index 708fd53..b81c7a5 100755 --- a/GTSRB/batch_slurm.sh +++ b/GTSRB/batch_slurm.sh @@ -9,7 +9,7 @@ # ----------------------------------------------- # # <!-- TITLE --> [GTSRB11] - SLURM batch script -# <!-- DESC --> Bash script for SLURM batch submission of GTSRB notebooks +# <!-- DESC --> Bash script for a Slurm batch submission of an ipython code # <!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) --> # # Soumission : sbatch /(...)/fidle/GTSRB/batch_slurm.sh diff --git a/IMDB/01-Embedding-Keras.ipynb b/IMDB/01-Embedding-Keras.ipynb index 275fb98..93627a9 100644 --- a/IMDB/01-Embedding-Keras.ipynb +++ b/IMDB/01-Embedding-Keras.ipynb @@ -6,8 +6,8 @@ "source": [ "<img width=\"800px\" src=\"../fidle/img/00-Fidle-header-01.svg\"></img>\n", "\n", - "# <!-- TITLE --> [IMDB1] - Text embedding with IMDB\n", - "<!-- DESC --> A very classical example of word embedding for text classification (sentiment analysis)\n", + "# <!-- TITLE --> [IMDB1] - Sentiment alalysis with text embedding\n", + "<!-- DESC --> A very classical example of word embedding with a dataset from Internet Movie Database (IMDB)\n", "<!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) -->\n", "\n", "## Objectives :\n", @@ -1021,7 +1021,7 @@ "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", - "version": "3.7.7" + "version": "3.7.9" } }, "nbformat": 4, diff --git a/IMDB/02-Prediction.ipynb b/IMDB/02-Prediction.ipynb index c8033a8..3536197 100644 --- a/IMDB/02-Prediction.ipynb +++ b/IMDB/02-Prediction.ipynb @@ -6,8 +6,8 @@ "source": [ "<img width=\"800px\" src=\"../fidle/img/00-Fidle-header-01.svg\"></img>\n", "\n", - "# <!-- TITLE --> [IMDB2] - Text embedding with IMDB - Reloaded\n", - "<!-- DESC --> Example of reusing a previously saved model\n", + "# <!-- TITLE --> [IMDB2] - Reload and reuse a saved model\n", + "<!-- DESC --> Retrieving a saved model to perform a sentiment analysis (movie review)\n", "<!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) -->\n", "\n", "## Objectives :\n", diff --git a/IMDB/03-LSTM-Keras.ipynb b/IMDB/03-LSTM-Keras.ipynb index 9f902e0..430dfd6 100644 --- a/IMDB/03-LSTM-Keras.ipynb +++ b/IMDB/03-LSTM-Keras.ipynb @@ -6,7 +6,7 @@ "source": [ "<img width=\"800px\" src=\"../fidle/img/00-Fidle-header-01.svg\"></img>\n", "\n", - "# <!-- TITLE --> [IMDB3] - Text embedding/LSTM model with IMDB\n", + "# <!-- TITLE --> [IMDB3] - Sentiment analysis with a LSTM network\n", "<!-- DESC --> Still the same problem, but with a network combining embedding and LSTM\n", "<!-- AUTHOR : Jean-Luc Parouty (CNRS/SIMaP) -->\n", "\n", @@ -980,7 +980,7 @@ "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", - "version": "3.7.7" + "version": "3.7.9" } }, "nbformat": 4, diff --git a/README.ipynb b/README.ipynb index b7d6380..2adbfbf 100644 --- a/README.ipynb +++ b/README.ipynb @@ -5,10 +5,10 @@ "execution_count": 1, "metadata": { "execution": { - "iopub.execute_input": "2021-01-08T10:27:08.480920Z", - "iopub.status.busy": "2021-01-08T10:27:08.480319Z", - "iopub.status.idle": "2021-01-08T10:27:08.483733Z", - "shell.execute_reply": "2021-01-08T10:27:08.483336Z" + "iopub.execute_input": "2021-01-08T21:53:24.359601Z", + "iopub.status.busy": "2021-01-08T21:53:24.359126Z", + "iopub.status.idle": "2021-01-08T21:53:24.367752Z", + "shell.execute_reply": "2021-01-08T21:53:24.368077Z" }, "jupyter": { "source_hidden": true @@ -85,31 +85,31 @@ "An example of classification using a dense neural network for the famous MNIST dataset\n", "\n", "### Images classification with Convolutional Neural Networks (CNN)\n", - "- **[GTSRB1](GTSRB/01-Preparation-of-data.ipynb)** - [CNN with GTSRB dataset - Data analysis and preparation](GTSRB/01-Preparation-of-data.ipynb) \n", - "Episode 1 : Data analysis and creation of a usable dataset\n", - "- **[GTSRB2](GTSRB/02-First-convolutions.ipynb)** - [CNN with GTSRB dataset - First convolutions](GTSRB/02-First-convolutions.ipynb) \n", - "Episode 2 : First convolutions and first results\n", - "- **[GTSRB3](GTSRB/03-Tracking-and-visualizing.ipynb)** - [CNN with GTSRB dataset - Monitoring ](GTSRB/03-Tracking-and-visualizing.ipynb) \n", - "Episode 3 : Monitoring and analysing training, managing checkpoints\n", - "- **[GTSRB4](GTSRB/04-Data-augmentation.ipynb)** - [CNN with GTSRB dataset - Data augmentation ](GTSRB/04-Data-augmentation.ipynb) \n", - "Episode 4 : Improving the results with data augmentation\n", - "- **[GTSRB5](GTSRB/05-Full-convolutions.ipynb)** - [CNN with GTSRB dataset - Full convolutions ](GTSRB/05-Full-convolutions.ipynb) \n", + "- **[GTSRB1](GTSRB/01-Preparation-of-data.ipynb)** - [Dataset analysis and preparation](GTSRB/01-Preparation-of-data.ipynb) \n", + "Episode 1 : Analysis of the GTSRB dataset and creation of an enhanced dataset\n", + "- **[GTSRB2](GTSRB/02-First-convolutions.ipynb)** - [First convolutions](GTSRB/02-First-convolutions.ipynb) \n", + "Episode 2 : First convolutions and first classification of our traffic signs\n", + "- **[GTSRB3](GTSRB/03-Tracking-and-visualizing.ipynb)** - [Training monitoring](GTSRB/03-Tracking-and-visualizing.ipynb) \n", + "Episode 3 : Monitoring, analysis and check points during a training session\n", + "- **[GTSRB4](GTSRB/04-Data-augmentation.ipynb)** - [Data augmentation ](GTSRB/04-Data-augmentation.ipynb) \n", + "Episode 4 : Adding data by data augmentation when we lack it, to improve our results\n", + "- **[GTSRB5](GTSRB/05-Full-convolutions.ipynb)** - [Full convolutions](GTSRB/05-Full-convolutions.ipynb) \n", "Episode 5 : A lot of models, a lot of datasets and a lot of results.\n", "- **[GTSRB6](GTSRB/06-Notebook-as-a-batch.ipynb)** - [Full convolutions as a batch](GTSRB/06-Notebook-as-a-batch.ipynb) \n", - "Episode 6 : Run Full convolution notebook as a batch\n", - "- **[GTSRB7](GTSRB/07-Show-report.ipynb)** - [CNN with GTSRB dataset - Show reports](GTSRB/07-Show-report.ipynb) \n", - "Episode 7 : Displaying a jobs report\n", - "- **[GTSRB10](GTSRB/batch_oar.sh)** - [OAR batch submission](GTSRB/batch_oar.sh) \n", - "Bash script for OAR batch submission of GTSRB notebook \n", + "Episode 6 : To compute bigger, use your notebook in batch mode\n", + "- **[GTSRB7](GTSRB/07-Show-report.ipynb)** - [Batch reportss](GTSRB/07-Show-report.ipynb) \n", + "Episode 7 : Displaying our jobs report, and the winner is...\n", + "- **[GTSRB10](GTSRB/batch_oar.sh)** - [OAR batch script submission](GTSRB/batch_oar.sh) \n", + "Bash script for an OAR batch submission of an ipython code\n", "- **[GTSRB11](GTSRB/batch_slurm.sh)** - [SLURM batch script](GTSRB/batch_slurm.sh) \n", - "Bash script for SLURM batch submission of GTSRB notebooks \n", + "Bash script for a Slurm batch submission of an ipython code\n", "\n", "### Sentiment analysis with word embedding\n", - "- **[IMDB1](IMDB/01-Embedding-Keras.ipynb)** - [Text embedding with IMDB](IMDB/01-Embedding-Keras.ipynb) \n", - "A very classical example of word embedding for text classification (sentiment analysis)\n", - "- **[IMDB2](IMDB/02-Prediction.ipynb)** - [Text embedding with IMDB - Reloaded](IMDB/02-Prediction.ipynb) \n", - "Example of reusing a previously saved model\n", - "- **[IMDB3](IMDB/03-LSTM-Keras.ipynb)** - [Text embedding/LSTM model with IMDB](IMDB/03-LSTM-Keras.ipynb) \n", + "- **[IMDB1](IMDB/01-Embedding-Keras.ipynb)** - [Sentiment alalysis with text embedding](IMDB/01-Embedding-Keras.ipynb) \n", + "A very classical example of word embedding with a dataset from Internet Movie Database (IMDB)\n", + "- **[IMDB2](IMDB/02-Prediction.ipynb)** - [Reload and reuse a saved model](IMDB/02-Prediction.ipynb) \n", + "Retrieving a saved model to perform a sentiment analysis (movie review)\n", + "- **[IMDB3](IMDB/03-LSTM-Keras.ipynb)** - [Sentiment analysis with a LSTM network](IMDB/03-LSTM-Keras.ipynb) \n", "Still the same problem, but with a network combining embedding and LSTM\n", "\n", "### Time series with Recurrent Neural Network (RNN)\n", @@ -187,6 +187,11 @@ } ], "metadata": { + "kernelspec": { + "display_name": "Python 3", + "language": "python", + "name": "python3" + }, "language_info": { "codemirror_mode": { "name": "ipython", diff --git a/README.md b/README.md index 57c65ab..457cc4e 100644 --- a/README.md +++ b/README.md @@ -65,31 +65,31 @@ A more advanced implementation of the precedent example An example of classification using a dense neural network for the famous MNIST dataset ### Images classification with Convolutional Neural Networks (CNN) -- **[GTSRB1](GTSRB/01-Preparation-of-data.ipynb)** - [CNN with GTSRB dataset - Data analysis and preparation](GTSRB/01-Preparation-of-data.ipynb) -Episode 1 : Data analysis and creation of a usable dataset -- **[GTSRB2](GTSRB/02-First-convolutions.ipynb)** - [CNN with GTSRB dataset - First convolutions](GTSRB/02-First-convolutions.ipynb) -Episode 2 : First convolutions and first results -- **[GTSRB3](GTSRB/03-Tracking-and-visualizing.ipynb)** - [CNN with GTSRB dataset - Monitoring ](GTSRB/03-Tracking-and-visualizing.ipynb) -Episode 3 : Monitoring and analysing training, managing checkpoints -- **[GTSRB4](GTSRB/04-Data-augmentation.ipynb)** - [CNN with GTSRB dataset - Data augmentation ](GTSRB/04-Data-augmentation.ipynb) -Episode 4 : Improving the results with data augmentation -- **[GTSRB5](GTSRB/05-Full-convolutions.ipynb)** - [CNN with GTSRB dataset - Full convolutions ](GTSRB/05-Full-convolutions.ipynb) +- **[GTSRB1](GTSRB/01-Preparation-of-data.ipynb)** - [Dataset analysis and preparation](GTSRB/01-Preparation-of-data.ipynb) +Episode 1 : Analysis of the GTSRB dataset and creation of an enhanced dataset +- **[GTSRB2](GTSRB/02-First-convolutions.ipynb)** - [First convolutions](GTSRB/02-First-convolutions.ipynb) +Episode 2 : First convolutions and first classification of our traffic signs +- **[GTSRB3](GTSRB/03-Tracking-and-visualizing.ipynb)** - [Training monitoring](GTSRB/03-Tracking-and-visualizing.ipynb) +Episode 3 : Monitoring, analysis and check points during a training session +- **[GTSRB4](GTSRB/04-Data-augmentation.ipynb)** - [Data augmentation ](GTSRB/04-Data-augmentation.ipynb) +Episode 4 : Adding data by data augmentation when we lack it, to improve our results +- **[GTSRB5](GTSRB/05-Full-convolutions.ipynb)** - [Full convolutions](GTSRB/05-Full-convolutions.ipynb) Episode 5 : A lot of models, a lot of datasets and a lot of results. - **[GTSRB6](GTSRB/06-Notebook-as-a-batch.ipynb)** - [Full convolutions as a batch](GTSRB/06-Notebook-as-a-batch.ipynb) -Episode 6 : Run Full convolution notebook as a batch -- **[GTSRB7](GTSRB/07-Show-report.ipynb)** - [CNN with GTSRB dataset - Show reports](GTSRB/07-Show-report.ipynb) -Episode 7 : Displaying a jobs report -- **[GTSRB10](GTSRB/batch_oar.sh)** - [OAR batch submission](GTSRB/batch_oar.sh) -Bash script for OAR batch submission of GTSRB notebook +Episode 6 : To compute bigger, use your notebook in batch mode +- **[GTSRB7](GTSRB/07-Show-report.ipynb)** - [Batch reportss](GTSRB/07-Show-report.ipynb) +Episode 7 : Displaying our jobs report, and the winner is... +- **[GTSRB10](GTSRB/batch_oar.sh)** - [OAR batch script submission](GTSRB/batch_oar.sh) +Bash script for an OAR batch submission of an ipython code - **[GTSRB11](GTSRB/batch_slurm.sh)** - [SLURM batch script](GTSRB/batch_slurm.sh) -Bash script for SLURM batch submission of GTSRB notebooks +Bash script for a Slurm batch submission of an ipython code ### Sentiment analysis with word embedding -- **[IMDB1](IMDB/01-Embedding-Keras.ipynb)** - [Text embedding with IMDB](IMDB/01-Embedding-Keras.ipynb) -A very classical example of word embedding for text classification (sentiment analysis) -- **[IMDB2](IMDB/02-Prediction.ipynb)** - [Text embedding with IMDB - Reloaded](IMDB/02-Prediction.ipynb) -Example of reusing a previously saved model -- **[IMDB3](IMDB/03-LSTM-Keras.ipynb)** - [Text embedding/LSTM model with IMDB](IMDB/03-LSTM-Keras.ipynb) +- **[IMDB1](IMDB/01-Embedding-Keras.ipynb)** - [Sentiment alalysis with text embedding](IMDB/01-Embedding-Keras.ipynb) +A very classical example of word embedding with a dataset from Internet Movie Database (IMDB) +- **[IMDB2](IMDB/02-Prediction.ipynb)** - [Reload and reuse a saved model](IMDB/02-Prediction.ipynb) +Retrieving a saved model to perform a sentiment analysis (movie review) +- **[IMDB3](IMDB/03-LSTM-Keras.ipynb)** - [Sentiment analysis with a LSTM network](IMDB/03-LSTM-Keras.ipynb) Still the same problem, but with a network combining embedding and LSTM ### Time series with Recurrent Neural Network (RNN) diff --git a/fidle/01 - Set and reset.ipynb b/fidle/01 - Set and reset.ipynb index 53a5a2f..7fbc4df 100644 --- a/fidle/01 - Set and reset.ipynb +++ b/fidle/01 - Set and reset.ipynb @@ -159,31 +159,31 @@ "An example of classification using a dense neural network for the famous MNIST dataset\n", "\n", "### Images classification with Convolutional Neural Networks (CNN)\n", - "- **[GTSRB1](GTSRB/01-Preparation-of-data.ipynb)** - [CNN with GTSRB dataset - Data analysis and preparation](GTSRB/01-Preparation-of-data.ipynb) \n", - "Episode 1 : Data analysis and creation of a usable dataset\n", - "- **[GTSRB2](GTSRB/02-First-convolutions.ipynb)** - [CNN with GTSRB dataset - First convolutions](GTSRB/02-First-convolutions.ipynb) \n", - "Episode 2 : First convolutions and first results\n", - "- **[GTSRB3](GTSRB/03-Tracking-and-visualizing.ipynb)** - [CNN with GTSRB dataset - Monitoring ](GTSRB/03-Tracking-and-visualizing.ipynb) \n", - "Episode 3 : Monitoring and analysing training, managing checkpoints\n", - "- **[GTSRB4](GTSRB/04-Data-augmentation.ipynb)** - [CNN with GTSRB dataset - Data augmentation ](GTSRB/04-Data-augmentation.ipynb) \n", - "Episode 4 : Improving the results with data augmentation\n", - "- **[GTSRB5](GTSRB/05-Full-convolutions.ipynb)** - [CNN with GTSRB dataset - Full convolutions ](GTSRB/05-Full-convolutions.ipynb) \n", + "- **[GTSRB1](GTSRB/01-Preparation-of-data.ipynb)** - [Dataset analysis and preparation](GTSRB/01-Preparation-of-data.ipynb) \n", + "Episode 1 : Analysis of the GTSRB dataset and creation of an enhanced dataset\n", + "- **[GTSRB2](GTSRB/02-First-convolutions.ipynb)** - [First convolutions](GTSRB/02-First-convolutions.ipynb) \n", + "Episode 2 : First convolutions and first classification of our traffic signs\n", + "- **[GTSRB3](GTSRB/03-Tracking-and-visualizing.ipynb)** - [Training monitoring](GTSRB/03-Tracking-and-visualizing.ipynb) \n", + "Episode 3 : Monitoring, analysis and check points during a training session\n", + "- **[GTSRB4](GTSRB/04-Data-augmentation.ipynb)** - [Data augmentation ](GTSRB/04-Data-augmentation.ipynb) \n", + "Episode 4 : Adding data by data augmentation when we lack it, to improve our results\n", + "- **[GTSRB5](GTSRB/05-Full-convolutions.ipynb)** - [Full convolutions](GTSRB/05-Full-convolutions.ipynb) \n", "Episode 5 : A lot of models, a lot of datasets and a lot of results.\n", "- **[GTSRB6](GTSRB/06-Notebook-as-a-batch.ipynb)** - [Full convolutions as a batch](GTSRB/06-Notebook-as-a-batch.ipynb) \n", - "Episode 6 : Run Full convolution notebook as a batch\n", - "- **[GTSRB7](GTSRB/07-Show-report.ipynb)** - [CNN with GTSRB dataset - Show reports](GTSRB/07-Show-report.ipynb) \n", - "Episode 7 : Displaying a jobs report\n", - "- **[GTSRB10](GTSRB/batch_oar.sh)** - [OAR batch submission](GTSRB/batch_oar.sh) \n", - "Bash script for OAR batch submission of GTSRB notebook \n", + "Episode 6 : To compute bigger, use your notebook in batch mode\n", + "- **[GTSRB7](GTSRB/07-Show-report.ipynb)** - [Batch reportss](GTSRB/07-Show-report.ipynb) \n", + "Episode 7 : Displaying our jobs report, and the winner is...\n", + "- **[GTSRB10](GTSRB/batch_oar.sh)** - [OAR batch script submission](GTSRB/batch_oar.sh) \n", + "Bash script for an OAR batch submission of an ipython code\n", "- **[GTSRB11](GTSRB/batch_slurm.sh)** - [SLURM batch script](GTSRB/batch_slurm.sh) \n", - "Bash script for SLURM batch submission of GTSRB notebooks \n", + "Bash script for a Slurm batch submission of an ipython code\n", "\n", "### Sentiment analysis with word embedding\n", - "- **[IMDB1](IMDB/01-Embedding-Keras.ipynb)** - [Text embedding with IMDB](IMDB/01-Embedding-Keras.ipynb) \n", - "A very classical example of word embedding for text classification (sentiment analysis)\n", - "- **[IMDB2](IMDB/02-Prediction.ipynb)** - [Text embedding with IMDB - Reloaded](IMDB/02-Prediction.ipynb) \n", - "Example of reusing a previously saved model\n", - "- **[IMDB3](IMDB/03-LSTM-Keras.ipynb)** - [Text embedding/LSTM model with IMDB](IMDB/03-LSTM-Keras.ipynb) \n", + "- **[IMDB1](IMDB/01-Embedding-Keras.ipynb)** - [Sentiment alalysis with text embedding](IMDB/01-Embedding-Keras.ipynb) \n", + "A very classical example of word embedding with a dataset from Internet Movie Database (IMDB)\n", + "- **[IMDB2](IMDB/02-Prediction.ipynb)** - [Reload and reuse a saved model](IMDB/02-Prediction.ipynb) \n", + "Retrieving a saved model to perform a sentiment analysis (movie review)\n", + "- **[IMDB3](IMDB/03-LSTM-Keras.ipynb)** - [Sentiment analysis with a LSTM network](IMDB/03-LSTM-Keras.ipynb) \n", "Still the same problem, but with a network combining embedding and LSTM\n", "\n", "### Time series with Recurrent Neural Network (RNN)\n", @@ -388,6 +388,10 @@ "new_cell['metadata']= { \"jupyter\": { \"source_hidden\": True} }\n", "notebook.cells.append(new_cell)\n", "\n", + "# --- Pour éviter une modification lors de l'ouverture du notebook\n", + "# pas génante, mais nécessite de resauvegarder le document à la fermeture...\n", + "notebook['metadata'][\"kernelspec\"] = {\"display_name\": \"Python 3\", \"language\": \"python\", \"name\": \"python3\" }\n", + "\n", "# ---- Run it\n", "#\n", "ep = ExecutePreprocessor(timeout=600, kernel_name=\"python3\")\n", diff --git a/fidle/log/catalog.json b/fidle/log/catalog.json index fbb2733..7d6b0ab 100644 --- a/fidle/log/catalog.json +++ b/fidle/log/catalog.json @@ -59,35 +59,35 @@ "id": "GTSRB1", "dirname": "GTSRB", "basename": "01-Preparation-of-data.ipynb", - "title": "CNN with GTSRB dataset - Data analysis and preparation", - "description": "Episode 1 : Data analysis and creation of a usable dataset" + "title": "Dataset analysis and preparation", + "description": "Episode 1 : Analysis of the GTSRB dataset and creation of an enhanced dataset" }, "GTSRB2": { "id": "GTSRB2", "dirname": "GTSRB", "basename": "02-First-convolutions.ipynb", - "title": "CNN with GTSRB dataset - First convolutions", - "description": "Episode 2 : First convolutions and first results" + "title": "First convolutions", + "description": "Episode 2 : First convolutions and first classification of our traffic signs" }, "GTSRB3": { "id": "GTSRB3", "dirname": "GTSRB", "basename": "03-Tracking-and-visualizing.ipynb", - "title": "CNN with GTSRB dataset - Monitoring ", - "description": "Episode 3 : Monitoring and analysing training, managing checkpoints" + "title": "Training monitoring", + "description": "Episode 3 : Monitoring, analysis and check points during a training session" }, "GTSRB4": { "id": "GTSRB4", "dirname": "GTSRB", "basename": "04-Data-augmentation.ipynb", - "title": "CNN with GTSRB dataset - Data augmentation ", - "description": "Episode 4 : Improving the results with data augmentation" + "title": "Data augmentation ", + "description": "Episode 4 : Adding data by data augmentation when we lack it, to improve our results" }, "GTSRB5": { "id": "GTSRB5", "dirname": "GTSRB", "basename": "05-Full-convolutions.ipynb", - "title": "CNN with GTSRB dataset - Full convolutions ", + "title": "Full convolutions", "description": "Episode 5 : A lot of models, a lot of datasets and a lot of results." }, "GTSRB6": { @@ -95,48 +95,48 @@ "dirname": "GTSRB", "basename": "06-Notebook-as-a-batch.ipynb", "title": "Full convolutions as a batch", - "description": "Episode 6 : Run Full convolution notebook as a batch" + "description": "Episode 6 : To compute bigger, use your notebook in batch mode" }, "GTSRB7": { "id": "GTSRB7", "dirname": "GTSRB", "basename": "07-Show-report.ipynb", - "title": "CNN with GTSRB dataset - Show reports", - "description": "Episode 7 : Displaying a jobs report" + "title": "Batch reportss", + "description": "Episode 7 : Displaying our jobs report, and the winner is..." }, "GTSRB10": { "id": "GTSRB10", "dirname": "GTSRB", "basename": "batch_oar.sh", - "title": "OAR batch submission", - "description": "Bash script for OAR batch submission of GTSRB notebook " + "title": "OAR batch script submission", + "description": "Bash script for an OAR batch submission of an ipython code" }, "GTSRB11": { "id": "GTSRB11", "dirname": "GTSRB", "basename": "batch_slurm.sh", "title": "SLURM batch script", - "description": "Bash script for SLURM batch submission of GTSRB notebooks " + "description": "Bash script for a Slurm batch submission of an ipython code" }, "IMDB1": { "id": "IMDB1", "dirname": "IMDB", "basename": "01-Embedding-Keras.ipynb", - "title": "Text embedding with IMDB", - "description": "A very classical example of word embedding for text classification (sentiment analysis)" + "title": "Sentiment alalysis with text embedding", + "description": "A very classical example of word embedding with a dataset from Internet Movie Database (IMDB)" }, "IMDB2": { "id": "IMDB2", "dirname": "IMDB", "basename": "02-Prediction.ipynb", - "title": "Text embedding with IMDB - Reloaded", - "description": "Example of reusing a previously saved model" + "title": "Reload and reuse a saved model", + "description": "Retrieving a saved model to perform a sentiment analysis (movie review)" }, "IMDB3": { "id": "IMDB3", "dirname": "IMDB", "basename": "03-LSTM-Keras.ipynb", - "title": "Text embedding/LSTM model with IMDB", + "title": "Sentiment analysis with a LSTM network", "description": "Still the same problem, but with a network combining embedding and LSTM" }, "SYNOP1": { -- GitLab