From 3acd23cf204d9b355069e108edef1c25a7097c96 Mon Sep 17 00:00:00 2001 From: Varuna Jayasiri Date: Sat, 27 Feb 2021 17:58:00 +0530 Subject: [PATCH] labml app links --- labml_nn/__init__.py | 2 +- labml_nn/transformers/compressive/readme.md | 2 +- labml_nn/transformers/feedback/README.md | 2 +- labml_nn/transformers/feedback/__init__.py | 2 +- labml_nn/transformers/feedback/experiment.py | 2 +- labml_nn/transformers/glu_variants/simple.py | 2 +- labml_nn/transformers/gpt/__init__.py | 2 +- labml_nn/transformers/switch/__init__.py | 2 +- labml_nn/transformers/switch/readme.md | 2 +- labml_nn/transformers/xl/__init__.py | 2 +- labml_nn/transformers/xl/readme.md | 2 +- readme.md | 2 +- 12 files changed, 12 insertions(+), 12 deletions(-) diff --git a/labml_nn/__init__.py b/labml_nn/__init__.py index ba60447c..f9d9e67e 100644 --- a/labml_nn/__init__.py +++ b/labml_nn/__init__.py @@ -1,5 +1,5 @@ """ -# [LabML Neural Networks](index.html) +# [labml.ai Neural Networks](index.html) This is a collection of simple PyTorch implementations of neural networks and related algorithms. diff --git a/labml_nn/transformers/compressive/readme.md b/labml_nn/transformers/compressive/readme.md index 71c3c7f3..0d154273 100644 --- a/labml_nn/transformers/compressive/readme.md +++ b/labml_nn/transformers/compressive/readme.md @@ -40,4 +40,4 @@ Here are [the training code](https://nn.labml.ai/transformers/compressive/experi model on the Tiny Shakespeare dataset. [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/compressive/experiment.ipynb) -[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://web.lab-ml.com/run?uuid=0d9b5338726c11ebb7c80242ac1c0002) +[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://app.labml.ai/run/0d9b5338726c11ebb7c80242ac1c0002) diff --git a/labml_nn/transformers/feedback/README.md b/labml_nn/transformers/feedback/README.md index b008214b..12d2ca85 100644 --- a/labml_nn/transformers/feedback/README.md +++ b/labml_nn/transformers/feedback/README.md @@ -32,4 +32,4 @@ Here's [the training code](experiment.html) and a notebook for training a feedba [Colab Notebook](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/feedback/experiment.ipynb) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/feedback/experiment.ipynb) -[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://web.lab-ml.com/run?uuid=d8eb9416530a11eb8fb50242ac1c0002) +[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://app.labml.ai/run/d8eb9416530a11eb8fb50242ac1c0002) diff --git a/labml_nn/transformers/feedback/__init__.py b/labml_nn/transformers/feedback/__init__.py index c99ed003..137504ea 100644 --- a/labml_nn/transformers/feedback/__init__.py +++ b/labml_nn/transformers/feedback/__init__.py @@ -37,7 +37,7 @@ We implemented a custom PyTorch function to improve performance. Here's [the training code](experiment.html) and a notebook for training a feedback transformer on Tiny Shakespeare dataset. [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/feedback/experiment.ipynb) -[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://web.lab-ml.com/run?uuid=d8eb9416530a11eb8fb50242ac1c0002) +[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://app.labml.ai/run/d8eb9416530a11eb8fb50242ac1c0002) """ import math diff --git a/labml_nn/transformers/feedback/experiment.py b/labml_nn/transformers/feedback/experiment.py index fe068372..b2eeafa6 100644 --- a/labml_nn/transformers/feedback/experiment.py +++ b/labml_nn/transformers/feedback/experiment.py @@ -13,7 +13,7 @@ where the keys and values are precalculated. Here's a Colab notebook for training a feedback transformer on Tiny Shakespeare dataset. [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/feedback/experiment.ipynb) -[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://web.lab-ml.com/run?uuid=d8eb9416530a11eb8fb50242ac1c0002) +[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://app.labml.ai/run/d8eb9416530a11eb8fb50242ac1c0002) """ import torch diff --git a/labml_nn/transformers/glu_variants/simple.py b/labml_nn/transformers/glu_variants/simple.py index 0ec82eef..10c801e9 100644 --- a/labml_nn/transformers/glu_variants/simple.py +++ b/labml_nn/transformers/glu_variants/simple.py @@ -15,7 +15,7 @@ We try different variants for the [position-wise feedforward network](../feed_fo We decided to write a simpler implementation to make it easier for readers who are not familiar.* [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/glu_variants/simple.ipynb) -[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://web.lab-ml.com/run?uuid=86b773f65fc911ebb2ac0242ac1c0002) +[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://app.labml.ai/run/86b773f65fc911ebb2ac0242ac1c0002) """ import dataclasses diff --git a/labml_nn/transformers/gpt/__init__.py b/labml_nn/transformers/gpt/__init__.py index cc083055..2e6a08d5 100644 --- a/labml_nn/transformers/gpt/__init__.py +++ b/labml_nn/transformers/gpt/__init__.py @@ -29,7 +29,7 @@ For the transformer we reuse the Here's a notebook for training a GPT model on Tiny Shakespeare dataset. [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/gpt/experiment.ipynb) -[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://web.lab-ml.com/run?uuid=0324c6d0562111eba65d0242ac1c0002) +[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://app.labml.ai/run/0324c6d0562111eba65d0242ac1c0002) """ import torch diff --git a/labml_nn/transformers/switch/__init__.py b/labml_nn/transformers/switch/__init__.py index 598c80be..4ebbd502 100644 --- a/labml_nn/transformers/switch/__init__.py +++ b/labml_nn/transformers/switch/__init__.py @@ -34,7 +34,7 @@ discusses dropping tokens when routing is not balanced. Here's [the training code](experiment.html) and a notebook for training a switch transformer on Tiny Shakespeare dataset. [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/switch/experiment.ipynb) -[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://web.lab-ml.com/run?uuid=c4656c605b9311eba13d0242ac1c0002) +[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://app.labml.ai/run/c4656c605b9311eba13d0242ac1c0002) """ import torch diff --git a/labml_nn/transformers/switch/readme.md b/labml_nn/transformers/switch/readme.md index 0ec13108..2f47406e 100644 --- a/labml_nn/transformers/switch/readme.md +++ b/labml_nn/transformers/switch/readme.md @@ -27,4 +27,4 @@ discusses dropping tokens when routing is not balanced. Here's [the training code](experiment.html) and a notebook for training a switch transformer on Tiny Shakespeare dataset. [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/switch/experiment.ipynb) -[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://web.lab-ml.com/run?uuid=c4656c605b9311eba13d0242ac1c0002) +[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://app.labml.ai/run/c4656c605b9311eba13d0242ac1c0002) diff --git a/labml_nn/transformers/xl/__init__.py b/labml_nn/transformers/xl/__init__.py index 54394961..b80b2cca 100644 --- a/labml_nn/transformers/xl/__init__.py +++ b/labml_nn/transformers/xl/__init__.py @@ -29,7 +29,7 @@ Annotated implementation of relative multi-headed attention is in [`relative_mha Here's [the training code](experiment.html) and a notebook for training a transformer XL model on Tiny Shakespeare dataset. [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/xl/experiment.ipynb) -[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://web.lab-ml.com/run?uuid=d3b6760c692e11ebb6a70242ac1c0002) +[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://app.labml.ai/run/d3b6760c692e11ebb6a70242ac1c0002) """ diff --git a/labml_nn/transformers/xl/readme.md b/labml_nn/transformers/xl/readme.md index f3d2436d..c10342c5 100644 --- a/labml_nn/transformers/xl/readme.md +++ b/labml_nn/transformers/xl/readme.md @@ -21,4 +21,4 @@ Annotated implementation of relative multi-headed attention is in [`relative_mha Here's [the training code](https://nn.labml.ai/transformers/xl/experiment.html) and a notebook for training a transformer XL model on Tiny Shakespeare dataset. [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lab-ml/nn/blob/master/labml_nn/transformers/xl/experiment.ipynb) -[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://web.lab-ml.com/run?uuid=d3b6760c692e11ebb6a70242ac1c0002) +[![View Run](https://img.shields.io/badge/labml-experiment-brightgreen)](https://app.labml.ai/run/d3b6760c692e11ebb6a70242ac1c0002) diff --git a/readme.md b/readme.md index 84422dd1..5cebe2b8 100644 --- a/readme.md +++ b/readme.md @@ -1,7 +1,7 @@ [![Join Slack](https://img.shields.io/badge/slack-chat-green.svg?logo=slack)](https://join.slack.com/t/labforml/shared_invite/zt-egj9zvq9-Dl3hhZqobexgT7aVKnD14g/) [![Twitter](https://img.shields.io/twitter/follow/labmlai?style=social)](https://twitter.com/labmlai) -# [LabML Neural Networks](https://nn.labml.ai/index.html) +# [labml.ai Neural Networks](https://nn.labml.ai/index.html) This is a collection of simple PyTorch implementations of neural networks and related algorithms.