From 52f3822d41150d66d963a2fdeb3101bf119e8713 Mon Sep 17 00:00:00 2001 From: Ekaterina Aidova Date: Fri, 29 Mar 2024 14:51:46 +0400 Subject: [PATCH] update links (#1867) --- .ci/check_links.py | 2 +- .github/workflows/docker.yml | 4 ++-- .github/workflows/generate_tags.yml | 2 +- .github/workflows/gh_pages_deploy.yml | 2 +- .github/workflows/install_requirements.yml | 2 +- .github/workflows/pip_conflicts_check.yml | 5 ++--- .github/workflows/spellcheck.yml | 4 ++-- .github/workflows/treon_precommit.yml | 8 ++------ README.md | 2 +- README_cn.md | 2 +- .../3D-pose-estimation-webcam/3D-pose-estimation.ipynb | 2 +- .../3D-segmentation-point-clouds.ipynb | 2 +- notebooks/3D-segmentation-point-clouds/README.md | 2 +- notebooks/amused-lightweight-text-to-image/README.md | 2 +- .../amused-lightweight-text-to-image.ipynb | 2 +- notebooks/async-api/README.md | 2 +- notebooks/attention-center/README.md | 2 +- notebooks/auto-device/README.md | 2 +- notebooks/bark-text-to-audio/bark-text-to-audio.ipynb | 2 +- notebooks/big-transfer-quantization/README.md | 2 +- .../blip-diffusion-subject-generation.ipynb | 2 +- notebooks/clip-language-saliency-map/README.md | 2 +- notebooks/convert-to-openvino/README.md | 2 +- notebooks/cross-lingual-books-alignment/README.md | 2 +- .../cross-lingual-books-alignment.ipynb | 2 +- .../decidiffusion-image-generation.ipynb | 2 +- notebooks/depth-anything/README.md | 2 +- notebooks/depth-anything/depth-anything.ipynb | 2 +- notebooks/depth-estimation-videpth/README.md | 2 +- .../depth-estimation-videpth.ipynb | 4 ++-- notebooks/detectron2-to-openvino/README.md | 2 +- notebooks/distilbert-sequence-classification/README.md | 2 +- .../distilbert-sequence-classification.ipynb | 2 +- notebooks/fast-segment-anything/README.md | 2 +- .../fast-segment-anything/fast-segment-anything.ipynb | 2 +- notebooks/grounded-segment-anything/README.md | 2 +- notebooks/handwritten-ocr/README.md | 2 +- notebooks/handwritten-ocr/handwritten-ocr.ipynb | 2 +- notebooks/hello-detection/README.md | 2 +- notebooks/hello-segmentation/README.md | 2 +- notebooks/hello-world/README.md | 2 +- notebooks/hugging-face-hub/README.md | 2 +- notebooks/image-classification-quantization/README.md | 2 +- .../instruct-pix2pix-image-editing.ipynb | 2 +- notebooks/knowledge-graphs-conve/README.md | 2 +- notebooks/language-quantize-bert/README.md | 2 +- .../latent-consistency-models-image-generation.ipynb | 2 +- .../latent-consistency-models-optimum-demo.ipynb | 2 +- .../llava-multimodal-chatbot.ipynb | 2 +- notebooks/machine-translation/README.md | 2 +- notebooks/magika-content-type-recognition/README.md | 2 +- notebooks/mobileclip-video-search/README.md | 2 +- notebooks/model-tools/README.md | 2 +- notebooks/music-generation/README.md | 2 +- notebooks/music-generation/music-generation.ipynb | 2 +- notebooks/named-entity-recognition/README.md | 2 +- notebooks/object-detection-webcam/README.md | 2 +- .../oneformer-segmentation/oneformer-segmentation.ipynb | 2 +- notebooks/openvino-api/README.md | 2 +- notebooks/openvino-tokenizers/README.md | 2 +- notebooks/openvoice/README.md | 2 +- notebooks/optical-character-recognition/README.md | 2 +- notebooks/optimize-preprocessing/README.md | 2 +- notebooks/paddle-ocr-webcam/README.md | 2 +- notebooks/paddle-ocr-webcam/paddle-ocr-webcam.ipynb | 2 +- notebooks/paddle-to-openvino/README.md | 2 +- notebooks/paint-by-example/paint-by-example.ipynb | 2 +- notebooks/performance-tricks/README.md | 4 ++-- notebooks/person-tracking-webcam/README.md | 2 +- notebooks/person-tracking-webcam/person-tracking.ipynb | 2 +- notebooks/pix2struct-docvqa/README.md | 2 +- .../pyannote-speaker-diarization.ipynb | 2 +- notebooks/pytorch-to-openvino/README.md | 4 ++-- notebooks/qrcode-monster/qrcode-monster.ipynb | 2 +- notebooks/question-answering/README.md | 2 +- .../riffusion-text-to-music/riffusion-text-to-music.ipynb | 2 +- notebooks/sdxl-turbo/sdxl-turbo.ipynb | 2 +- notebooks/segmenter-semantic-segmentation/README.md | 2 +- notebooks/siglip-zero-shot-image-classification/README.md | 2 +- .../siglip-zero-shot-image-classification.ipynb | 2 +- notebooks/softvc-voice-conversion/README.md | 2 +- .../sound-generation-audioldm2.ipynb | 2 +- notebooks/sparsity-optimization/README.md | 2 +- notebooks/speculative-sampling/README.md | 2 +- notebooks/speech-recognition-quantization/README.md | 2 +- notebooks/speech-to-text/README.md | 2 +- .../stable-diffusion-ip-adapter.ipynb | 2 +- .../stable-diffusion-keras-cv.ipynb | 2 +- .../stable-diffusion-torchdynamo-backend.ipynb | 2 +- .../stable-diffusion-v2-infinite-zoom.ipynb | 2 +- .../stable-diffusion-v2-optimum-demo.ipynb | 2 +- .../stable-diffusion-v2-text-to-image-demo.ipynb | 2 +- .../stable-diffusion-v2-text-to-image.ipynb | 2 +- notebooks/stable-diffusion-xl/segmind-vegart.ipynb | 2 +- notebooks/stable-diffusion-xl/ssd-b1.ipynb | 2 +- notebooks/stable-diffusion-xl/stable-diffusion-xl.ipynb | 2 +- notebooks/style-transfer-webcam/README.md | 2 +- notebooks/style-transfer-webcam/style-transfer.ipynb | 2 +- notebooks/surya-line-level-text-detection/README.md | 2 +- .../surya-line-level-text-detection.ipynb | 2 +- notebooks/table-question-answering/README.md | 2 +- notebooks/tensorflow-classification-to-openvino/README.md | 2 +- notebooks/tensorflow-hub/README.md | 2 +- notebooks/tensorflow-hub/tensorflow-hub.ipynb | 2 +- .../tensorflow-object-detection-to-openvino/README.md | 4 ++-- .../tensorflow-quantization-aware-training/README.md | 2 +- .../tensorflow-training-openvino-nncf.ipynb | 2 +- notebooks/text-prediction/README.md | 2 +- notebooks/tflite-selfie-segmentation/README.md | 2 +- notebooks/tflite-to-openvino/README.md | 2 +- notebooks/tiny-sd-image-generation/README.md | 2 +- .../tiny-sd-image-generation.ipynb | 2 +- notebooks/typo-detector/typo-detector.ipynb | 2 +- notebooks/vision-background-removal/README.md | 2 +- .../vision-background-removal.ipynb | 2 +- notebooks/vision-monodepth/README.md | 2 +- notebooks/vision-paddlegan-anime/README.md | 2 +- notebooks/vision-paddlegan-superresolution/README.md | 2 +- .../vision-paddlegan-superresolution.ipynb | 2 +- notebooks/vision-superresolution/README.md | 2 +- .../vision-superresolution-image.ipynb | 2 +- .../vision-superresolution-video.ipynb | 2 +- notebooks/whisper-subtitles-generation/README.md | 2 +- .../wuerstchen-image-generation.ipynb | 2 +- notebooks/yolov8-optimization/README.md | 8 ++++---- notebooks/zeroscope-text2video/zeroscope-text2video.ipynb | 2 +- .../src/notebook-metadata/notebook-metadata-collector.js | 2 +- 127 files changed, 138 insertions(+), 143 deletions(-) diff --git a/.ci/check_links.py b/.ci/check_links.py index 48cdf53bc15..d821488a115 100644 --- a/.ci/check_links.py +++ b/.ci/check_links.py @@ -36,7 +36,7 @@ def get_all_references_from_md(md_path): def validate_colab_url(url: str) -> bool: - OPENVINO_COLAB_URL_PREFIX = 'https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/' + OPENVINO_COLAB_URL_PREFIX = 'https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/' if not url.startswith(OPENVINO_COLAB_URL_PREFIX): return diff --git a/.github/workflows/docker.yml b/.github/workflows/docker.yml index 5290efc97a5..e468918e902 100644 --- a/.github/workflows/docker.yml +++ b/.github/workflows/docker.yml @@ -10,7 +10,7 @@ on: push: branches: - 'main' - - 'master' + - 'latest' paths: - 'Dockerfile' - '.docker/**' @@ -23,7 +23,7 @@ on: pull_request: branches: - 'main' - - 'master' + - 'latest' - 'develop' paths: - 'Dockerfile' diff --git a/.github/workflows/generate_tags.yml b/.github/workflows/generate_tags.yml index dddf2440790..6bd25cf113d 100644 --- a/.github/workflows/generate_tags.yml +++ b/.github/workflows/generate_tags.yml @@ -6,7 +6,7 @@ on: pull_request: branches: - 'main' - - 'master' + - 'latest' paths: - 'notebooks/**.ipynb' - '.ci/keywords.json' diff --git a/.github/workflows/gh_pages_deploy.yml b/.github/workflows/gh_pages_deploy.yml index 3fea7546c2b..5253ac11fe0 100644 --- a/.github/workflows/gh_pages_deploy.yml +++ b/.github/workflows/gh_pages_deploy.yml @@ -4,7 +4,7 @@ on: push: branches: - 'main' - - 'master' + - 'latest' workflow_dispatch: concurrency: diff --git a/.github/workflows/install_requirements.yml b/.github/workflows/install_requirements.yml index e0a934def61..06e2d9e19be 100644 --- a/.github/workflows/install_requirements.yml +++ b/.github/workflows/install_requirements.yml @@ -5,7 +5,7 @@ on: pull_request: branches: - 'main' - - 'master' + - 'latest' paths: - '.github/workflows/install_requirements.yml' - 'check_install.py' diff --git a/.github/workflows/pip_conflicts_check.yml b/.github/workflows/pip_conflicts_check.yml index e6e72f23a93..314a1879e30 100644 --- a/.github/workflows/pip_conflicts_check.yml +++ b/.github/workflows/pip_conflicts_check.yml @@ -5,14 +5,13 @@ on: pull_request: branches: - 'main' - - 'master' - - 'develop' + - 'latest' paths: - 'notebooks/**.ipynb' push: branches: - 'main' - - 'master' + - 'latest' paths: - 'notebooks/**.ipynb' diff --git a/.github/workflows/spellcheck.yml b/.github/workflows/spellcheck.yml index 5599e43f3cd..db26e513565 100644 --- a/.github/workflows/spellcheck.yml +++ b/.github/workflows/spellcheck.yml @@ -5,7 +5,7 @@ on: pull_request: branches: - 'main' - - 'master' + - 'latest' - 'develop' paths: - '.ci/spellcheck/*' @@ -16,7 +16,7 @@ on: push: branches: - 'main' - - 'master' + - 'latest' paths: - '.ci/spellcheck/*' - 'README.md' diff --git a/.github/workflows/treon_precommit.yml b/.github/workflows/treon_precommit.yml index f8a45eb43c4..b38c0161090 100644 --- a/.github/workflows/treon_precommit.yml +++ b/.github/workflows/treon_precommit.yml @@ -5,9 +5,7 @@ on: pull_request: branches: - 'main' - - 'master' - - '2021.4' - - 'develop' + - 'latest' paths: - 'notebooks/**.ipynb' - 'notebooks/**.py' @@ -18,9 +16,7 @@ on: push: branches: - 'main' - - 'master' - - '2021.4' - - 'develop' + - 'latest' paths: - 'notebooks/**.ipynb' - 'notebooks/**.py' diff --git a/README.md b/README.md index e8b2771ffa1..9a8ea988433 100644 --- a/README.md +++ b/README.md @@ -2,7 +2,7 @@ English | [简体中文](README_cn.md)

📚 OpenVINO™ Notebooks

-[![Apache License Version 2.0](https://img.shields.io/badge/license-Apache_2.0-green.svg)](https://github.com/openvinotoolkit/openvino_notebooks/blob/master/LICENSE) +[![Apache License Version 2.0](https://img.shields.io/badge/license-Apache_2.0-green.svg)](https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/LICENSE) [![CI](https://github.com/openvinotoolkit/openvino_notebooks/actions/workflows/treon_precommit.yml/badge.svg?event=push)](https://github.com/openvinotoolkit/openvino_notebooks/actions/workflows/treon_precommit.yml?query=event%3Apush) [![CI](https://github.com/openvinotoolkit/openvino_notebooks/actions/workflows/docker.yml/badge.svg?event=push)](https://github.com/openvinotoolkit/openvino_notebooks/actions/workflows/docker.yml?query=event%3Apush) diff --git a/README_cn.md b/README_cn.md index 68aa581a9a7..2db46768304 100644 --- a/README_cn.md +++ b/README_cn.md @@ -2,7 +2,7 @@

📚 OpenVINO™ Notebooks

-[![Apache License Version 2.0](https://img.shields.io/badge/license-Apache_2.0-green.svg)](https://github.com/openvinotoolkit/openvino_notebooks/blob/master/LICENSE) +[![Apache License Version 2.0](https://img.shields.io/badge/license-Apache_2.0-green.svg)](https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/LICENSE) [![CI](https://github.com/openvinotoolkit/openvino_notebooks/actions/workflows/treon_precommit.yml/badge.svg?event=push)](https://github.com/openvinotoolkit/openvino_notebooks/actions/workflows/treon_precommit.yml?query=event%3Apush) [![CI](https://github.com/openvinotoolkit/openvino_notebooks/actions/workflows/docker.yml/badge.svg?event=push)](https://github.com/openvinotoolkit/openvino_notebooks/actions/workflows/docker.yml?query=event%3Apush) diff --git a/notebooks/3D-pose-estimation-webcam/3D-pose-estimation.ipynb b/notebooks/3D-pose-estimation-webcam/3D-pose-estimation.ipynb index aab0d065361..da44b9a22e4 100644 --- a/notebooks/3D-pose-estimation-webcam/3D-pose-estimation.ipynb +++ b/notebooks/3D-pose-estimation-webcam/3D-pose-estimation.ipynb @@ -626,7 +626,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/3D-pose-estimation-webcam/3D-pose-estimation.gif?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/3D-pose-estimation-webcam/3D-pose-estimation.gif?raw=true", "tags": { "categories": [ "Live Demos" diff --git a/notebooks/3D-segmentation-point-clouds/3D-segmentation-point-clouds.ipynb b/notebooks/3D-segmentation-point-clouds/3D-segmentation-point-clouds.ipynb index 63d30abe550..6ced333b126 100644 --- a/notebooks/3D-segmentation-point-clouds/3D-segmentation-point-clouds.ipynb +++ b/notebooks/3D-segmentation-point-clouds/3D-segmentation-point-clouds.ipynb @@ -413,7 +413,7 @@ "version": "3.11.5" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/3D-segmentation-point-clouds/3D-segmentation-point-clouds.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/3D-segmentation-point-clouds/3D-segmentation-point-clouds.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/3D-segmentation-point-clouds/README.md b/notebooks/3D-segmentation-point-clouds/README.md index bc9f7512373..c4c098c75b2 100644 --- a/notebooks/3D-segmentation-point-clouds/README.md +++ b/notebooks/3D-segmentation-point-clouds/README.md @@ -1,7 +1,7 @@ # Part Segmentation of 3D Point Clouds with OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/openvinotoolkit/openvino_notebooks/HEAD?labpath=notebooks%2F3D-segmentation-point-clouds%2F3D-segmentation-point-clouds.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/3D-segmentation-point-clouds/3D-segmentation-point-clouds.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/3D-segmentation-point-clouds/3D-segmentation-point-clouds.ipynb)

diff --git a/notebooks/amused-lightweight-text-to-image/README.md b/notebooks/amused-lightweight-text-to-image/README.md index 33862090836..06d57f96003 100644 --- a/notebooks/amused-lightweight-text-to-image/README.md +++ b/notebooks/amused-lightweight-text-to-image/README.md @@ -1,6 +1,6 @@ # Lightweight image generation with aMUSEd and OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/amused-lightweight-text-to-image/amused-lightweight-text-to-image.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/amused-lightweight-text-to-image/amused-lightweight-text-to-image.ipynb) diff --git a/notebooks/amused-lightweight-text-to-image/amused-lightweight-text-to-image.ipynb b/notebooks/amused-lightweight-text-to-image/amused-lightweight-text-to-image.ipynb index 82e924ffd9c..29417faa82a 100644 --- a/notebooks/amused-lightweight-text-to-image/amused-lightweight-text-to-image.ipynb +++ b/notebooks/amused-lightweight-text-to-image/amused-lightweight-text-to-image.ipynb @@ -693,7 +693,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/amused-lightweight-text-to-image/amused-lightweight-text-to-image.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/amused-lightweight-text-to-image/amused-lightweight-text-to-image.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/async-api/README.md b/notebooks/async-api/README.md index 603f8ae5fe0..a453c28dd07 100644 --- a/notebooks/async-api/README.md +++ b/notebooks/async-api/README.md @@ -1,7 +1,7 @@ # Asynchronous Inference with OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/openvinotoolkit/openvino_notebooks/HEAD?labpath=notebooks%2Fasync-api%2Fasync-api.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/async-api/async-api.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/async-api/async-api.ipynb) This notebook demonstrates how to use the [Async API](https://docs.openvino.ai/2024/openvino-workflow/running-inference/optimize-inference/general-optimizations.html) and [`AsyncInferQueue`](https://docs.openvino.ai/2024/openvino-workflow/running-inference/integrate-openvino-with-your-application/python-api-exclusives.html#asyncinferqueue) for asynchronous execution with OpenVINO. diff --git a/notebooks/attention-center/README.md b/notebooks/attention-center/README.md index 254ac26df32..88f65da5f6a 100644 --- a/notebooks/attention-center/README.md +++ b/notebooks/attention-center/README.md @@ -1,6 +1,6 @@ # The attention center model with OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/attention-center/attention-center.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/attention-center/attention-center.ipynb) This notebook demonstrates how the use [attention center model](https://github.com/google/attention-center/tree/main) with OpenVINO. This model is in the [TensorFlow Lite format](https://www.tensorflow.org/lite), which is supported in OpenVINO now by TFLite frontend. Check out [this article](https://opensource.googleblog.com/2022/12/open-sourcing-attention-center-model.html) to find more information about this model. The attention-center model takes an RGB image as input and return a 2D point as result, which is the predicted center of human attention on the image. diff --git a/notebooks/auto-device/README.md b/notebooks/auto-device/README.md index 158095e1f25..463df75d9f7 100644 --- a/notebooks/auto-device/README.md +++ b/notebooks/auto-device/README.md @@ -1,7 +1,7 @@ # Introduction to Auto Device Selection in OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fauto-device%2Fauto-device.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/auto-device/auto-device.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/auto-device/auto-device.ipynb) This notebook shows how to do inference with Automatic Device Selection (AUTO). To learn more about the logic of this mode, refer to the [Automatic device selection](https://docs.openvino.ai/2024/openvino-workflow/running-inference/inference-devices-and-modes/auto-device-selection.html) article. diff --git a/notebooks/bark-text-to-audio/bark-text-to-audio.ipynb b/notebooks/bark-text-to-audio/bark-text-to-audio.ipynb index 0d990c0e8bd..fec1f425d3b 100644 --- a/notebooks/bark-text-to-audio/bark-text-to-audio.ipynb +++ b/notebooks/bark-text-to-audio/bark-text-to-audio.ipynb @@ -1373,7 +1373,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/bark-text-to-audio/bark-text-to-audio.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/bark-text-to-audio/bark-text-to-audio.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/big-transfer-quantization/README.md b/notebooks/big-transfer-quantization/README.md index 751373097ab..5b6a3d72d69 100644 --- a/notebooks/big-transfer-quantization/README.md +++ b/notebooks/big-transfer-quantization/README.md @@ -18,4 +18,4 @@ This tutorial consists of the following steps: ## Installation Instructions -If you have not installed all required dependencies, follow the [Installation Guide](https://github.com/openvinotoolkit/openvino_notebooks/blob/master/README.md#-installation-guide) \ No newline at end of file +If you have not installed all required dependencies, follow the [Installation Guide](https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/README.md#-installation-guide) \ No newline at end of file diff --git a/notebooks/blip-diffusion-subject-generation/blip-diffusion-subject-generation.ipynb b/notebooks/blip-diffusion-subject-generation/blip-diffusion-subject-generation.ipynb index 30373e5528b..4838a161a18 100644 --- a/notebooks/blip-diffusion-subject-generation/blip-diffusion-subject-generation.ipynb +++ b/notebooks/blip-diffusion-subject-generation/blip-diffusion-subject-generation.ipynb @@ -1967,7 +1967,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/blip-diffusion-subject-generation/blip-diffusion-subject-generation.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/blip-diffusion-subject-generation/blip-diffusion-subject-generation.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/clip-language-saliency-map/README.md b/notebooks/clip-language-saliency-map/README.md index 30cfaed1cd6..fb9d00b4c32 100644 --- a/notebooks/clip-language-saliency-map/README.md +++ b/notebooks/clip-language-saliency-map/README.md @@ -1,5 +1,5 @@ # Language-Visual Saliency with CLIP and OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/clip-language-saliency-map/clip-language-saliency-map.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/clip-language-saliency-map/clip-language-saliency-map.ipynb) The notebook will cover the following topics: diff --git a/notebooks/convert-to-openvino/README.md b/notebooks/convert-to-openvino/README.md index 0f0feb76d52..34908b77662 100644 --- a/notebooks/convert-to-openvino/README.md +++ b/notebooks/convert-to-openvino/README.md @@ -1,7 +1,7 @@ # OpenVINO™ model conversion API tutorial [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fconvert-to-openvino%2Fconvert-to-openvino.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/convert-to-openvino/convert-to-openvino.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/convert-to-openvino/convert-to-openvino.ipynb) This notebook shows how to convert a model from its original framework format to the OpenVINO IR. It describes the Python conversion API and the OpenVINO Model Converter command-line tool. It provides examples of converting Hugging Face and PyTorch models to OpenVINO IR. diff --git a/notebooks/cross-lingual-books-alignment/README.md b/notebooks/cross-lingual-books-alignment/README.md index 2b969b1789c..8f99b1bf24c 100644 --- a/notebooks/cross-lingual-books-alignment/README.md +++ b/notebooks/cross-lingual-books-alignment/README.md @@ -1,6 +1,6 @@ # Cross-lingual Books Alignment With Transformers and OpenVINO™ -[![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/openvinotoolkit/openvino_notebooks/HEAD?labpath=notebooks%2Fcross-lingual-books-alignment%2Fcross-lingual-books-alignment.ipynb) [![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/cross-lingual-books-alignment/cross-lingual-books-alignment.ipynb) +[![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/openvinotoolkit/openvino_notebooks/HEAD?labpath=notebooks%2Fcross-lingual-books-alignment%2Fcross-lingual-books-alignment.ipynb) [![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/cross-lingual-books-alignment/cross-lingual-books-alignment.ipynb) Cross-lingual text alignment is the task of matching sentences in a pair of texts that are translations of each other. In this notebook, you'll learn how to use a deep learning model to create a parallel book in English and German diff --git a/notebooks/cross-lingual-books-alignment/cross-lingual-books-alignment.ipynb b/notebooks/cross-lingual-books-alignment/cross-lingual-books-alignment.ipynb index be719e485c4..1228f2e562a 100644 --- a/notebooks/cross-lingual-books-alignment/cross-lingual-books-alignment.ipynb +++ b/notebooks/cross-lingual-books-alignment/cross-lingual-books-alignment.ipynb @@ -1385,7 +1385,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/cross-lingual-books-alignment/cross-lingual-books-alignment.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/cross-lingual-books-alignment/cross-lingual-books-alignment.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/decidiffusion-image-generation/decidiffusion-image-generation.ipynb b/notebooks/decidiffusion-image-generation/decidiffusion-image-generation.ipynb index 9eaacd6db80..a61de0290d3 100644 --- a/notebooks/decidiffusion-image-generation/decidiffusion-image-generation.ipynb +++ b/notebooks/decidiffusion-image-generation/decidiffusion-image-generation.ipynb @@ -1848,7 +1848,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/decidiffusion-image-generation/decidiffusion-image-generation.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/decidiffusion-image-generation/decidiffusion-image-generation.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/depth-anything/README.md b/notebooks/depth-anything/README.md index a9f78673030..31ae52a817a 100644 --- a/notebooks/depth-anything/README.md +++ b/notebooks/depth-anything/README.md @@ -1,7 +1,7 @@ # Depth estimation with DepthAnything and OpenVINO [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fdepth-anythingh%2Fdepth-anything.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/depth-anything/depth-anything.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/depth-anything/depth-anything.ipynb) ![depth_map.gif](https://github.com/openvinotoolkit/openvino_notebooks/assets/29454499/a9a16658-512f-470c-a33c-0e1f9d0ae72c) diff --git a/notebooks/depth-anything/depth-anything.ipynb b/notebooks/depth-anything/depth-anything.ipynb index 52040fbe07c..ea995bce3cc 100644 --- a/notebooks/depth-anything/depth-anything.ipynb +++ b/notebooks/depth-anything/depth-anything.ipynb @@ -1284,7 +1284,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/depth-anything/depth-anything.gif?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/depth-anything/depth-anything.gif?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/depth-estimation-videpth/README.md b/notebooks/depth-estimation-videpth/README.md index 6e0eca992c0..9c58a8456c0 100644 --- a/notebooks/depth-estimation-videpth/README.md +++ b/notebooks/depth-estimation-videpth/README.md @@ -31,4 +31,4 @@ The notebook contains a detailed tutorial of the visual-inertial depth estimatio ## Installation Instructions -If you have not installed all required dependencies, follow the [Installation Guide](https://github.com/pronoym99/openvino_notebooks/blob/master/README.md#-installation-guide). +If you have not installed all required dependencies, follow the [Installation Guide](https://github.com/pronoym99/openvino_notebooks/blob/latest/README.md#-installation-guide). diff --git a/notebooks/depth-estimation-videpth/depth-estimation-videpth.ipynb b/notebooks/depth-estimation-videpth/depth-estimation-videpth.ipynb index 7484e49379a..be6acfd93cb 100644 --- a/notebooks/depth-estimation-videpth/depth-estimation-videpth.ipynb +++ b/notebooks/depth-estimation-videpth/depth-estimation-videpth.ipynb @@ -1124,7 +1124,7 @@ "\n", "Now role of both the dummy inputs i.e. the dummy image as well as its associated depth map is now over. Since we have access to the compiled models now, we can load the *one* image available to us for pure inferencing purposes and run all the above steps one by one till plotting of the depth map.\n", "\n", - "If you haven't noticed already the data directory of this tutorial has been arranged as follows. This allows us to comply to these [rules](https://github.com/pronoym99/openvino_notebooks/blob/master/CONTRIBUTING.md#file-structure).\n", + "If you haven't noticed already the data directory of this tutorial has been arranged as follows. This allows us to comply to these [rules](https://github.com/pronoym99/openvino_notebooks/blob/latest/CONTRIBUTING.md#file-structure).\n", "\n", " ```bash\n", " data\n", @@ -1325,7 +1325,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/depth-estimation-videpth/depth-estimation-videpth.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/depth-estimation-videpth/depth-estimation-videpth.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/detectron2-to-openvino/README.md b/notebooks/detectron2-to-openvino/README.md index aaa64878bb8..356ff98c459 100644 --- a/notebooks/detectron2-to-openvino/README.md +++ b/notebooks/detectron2-to-openvino/README.md @@ -1,7 +1,7 @@ # Convert Detection2 Models to OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fdetectron2-to-openvino%2Fdetectron2-to-openvino.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/detectron2-to-openvino/detectron2-to-openvino.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/detectron2-to-openvino/detectron2-to-openvino.ipynb) diff --git a/notebooks/distilbert-sequence-classification/README.md b/notebooks/distilbert-sequence-classification/README.md index 444e94d808e..97f8e220140 100644 --- a/notebooks/distilbert-sequence-classification/README.md +++ b/notebooks/distilbert-sequence-classification/README.md @@ -2,7 +2,7 @@ Sequence Classification (or Text Classification) is the NLP task of predicting a label for a sequence of words. [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fdistilbert-sequence-classification%2Fdistilbert-sequence-classification.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/distilbert-sequence-classification/distilbert-sequence-classification.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/distilbert-sequence-classification/distilbert-sequence-classification.ipynb) Sentiment analysis is a sub-task of Sequence Classification. It is the use of natural language processing, text analysis, computational linguistics, and biometrics to systematically identify, extract, quantify, and study affective states and subjective information. This notebook performs sentiment analysis using OpenVINO. We will use the transformer-based [DistilBERT base uncased finetuned SST-2](https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english) model from Hugging Face converting to OpenVINO IR format using OpenVINO PyTorch Frontend and model conversion Python API. We can also replace the model with the other BERT-based models for sequence classification. The model predicts one of two classes: Positive or Negative, after analyzing the sentiment of any given text. The notebook also estimates time required for inference. diff --git a/notebooks/distilbert-sequence-classification/distilbert-sequence-classification.ipynb b/notebooks/distilbert-sequence-classification/distilbert-sequence-classification.ipynb index 3e020638f56..efc20bb1b73 100644 --- a/notebooks/distilbert-sequence-classification/distilbert-sequence-classification.ipynb +++ b/notebooks/distilbert-sequence-classification/distilbert-sequence-classification.ipynb @@ -430,7 +430,7 @@ "useGoogleTranslate": true }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/distilbert-sequence-classification/distilbert-sequence-classification.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/distilbert-sequence-classification/distilbert-sequence-classification.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/fast-segment-anything/README.md b/notebooks/fast-segment-anything/README.md index 627a83b42b0..eb8459ea6f1 100644 --- a/notebooks/fast-segment-anything/README.md +++ b/notebooks/fast-segment-anything/README.md @@ -1,7 +1,7 @@ # Object segmentations with FastSAM and OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Ffast-segment-anything%2Ffast-segment-anything.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/fast-segment-anything/fast-segment-anything.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/fast-segment-anything/fast-segment-anything.ipynb) The Fast Segment Anything Model (FastSAM) is a real-time CNN-based model that can segment any object within an image based on various user prompts. `Segment Anything` task is designed to make vision tasks easier by providing an efficient way to identify objects in an image. FastSAM significantly reduces computational demands while maintaining competitive performance, making it a practical choice for a variety of vision tasks. diff --git a/notebooks/fast-segment-anything/fast-segment-anything.ipynb b/notebooks/fast-segment-anything/fast-segment-anything.ipynb index 494f7ee65bb..d510b9b05c0 100644 --- a/notebooks/fast-segment-anything/fast-segment-anything.ipynb +++ b/notebooks/fast-segment-anything/fast-segment-anything.ipynb @@ -1088,7 +1088,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/fast-segment-anything/fast-segment-anything.gif?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/fast-segment-anything/fast-segment-anything.gif?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/grounded-segment-anything/README.md b/notebooks/grounded-segment-anything/README.md index 3a4bdb1f35c..64768085f65 100644 --- a/notebooks/grounded-segment-anything/README.md +++ b/notebooks/grounded-segment-anything/README.md @@ -1,6 +1,6 @@ # Object detection and masking from prompts with GroundedSAM (GroundingDINO + SAM) and OpenVINO -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/grounded-segment-anything/grounded-segment-anything.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/grounded-segment-anything/grounded-segment-anything.ipynb) In this notebook, we provide the OpenVINO™ optimization for the combination of GroundingDINO + SAM = [GroundedSAM](https://github.com/IDEA-Research/Grounded-Segment-Anything) on Intel® platforms. diff --git a/notebooks/handwritten-ocr/README.md b/notebooks/handwritten-ocr/README.md index 5d059ee7f90..c5aed6de743 100644 --- a/notebooks/handwritten-ocr/README.md +++ b/notebooks/handwritten-ocr/README.md @@ -1,7 +1,7 @@ # Handwritten OCR with OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fhandwritten-ocr%2Fhandwritten-ocr.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/handwritten-ocr/handwritten-ocr.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/handwritten-ocr/handwritten-ocr.ipynb) handwritten_simplified_chinese_test diff --git a/notebooks/handwritten-ocr/handwritten-ocr.ipynb b/notebooks/handwritten-ocr/handwritten-ocr.ipynb index 30117af1b81..f2aaa5d15b2 100644 --- a/notebooks/handwritten-ocr/handwritten-ocr.ipynb +++ b/notebooks/handwritten-ocr/handwritten-ocr.ipynb @@ -571,7 +571,7 @@ "version": "3.11.5" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/handwritten-ocr/handwritten-ocr.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/handwritten-ocr/handwritten-ocr.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/hello-detection/README.md b/notebooks/hello-detection/README.md index 0172995a02d..a8d8ea81584 100644 --- a/notebooks/hello-detection/README.md +++ b/notebooks/hello-detection/README.md @@ -2,7 +2,7 @@ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fhello-detection%2Fhello-detection.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/hello-detection/hello-detection.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/hello-detection/hello-detection.ipynb) | | | | --------------------------------------------------------------------------------------------------------------------------- | --------------------------------------------------------------------------------------------------------------------------- | diff --git a/notebooks/hello-segmentation/README.md b/notebooks/hello-segmentation/README.md index 8dcb963b120..1438be78f7e 100644 --- a/notebooks/hello-segmentation/README.md +++ b/notebooks/hello-segmentation/README.md @@ -1,7 +1,7 @@ # Introduction to Segmentation in OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fhello-segmentation%2Fhello-segmentation.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/hello-segmentation/hello-segmentation.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/hello-segmentation/hello-segmentation.ipynb) | | | | --------------------------------------------------------------------------------------------------------------------------- | --------------------------------------------------------------------------------------------------------------------------- | diff --git a/notebooks/hello-world/README.md b/notebooks/hello-world/README.md index eb9ed1ba715..66f27593265 100644 --- a/notebooks/hello-world/README.md +++ b/notebooks/hello-world/README.md @@ -1,7 +1,7 @@ # Introduction to OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fhello-world%2Fhello-world.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/hello-world/hello-world.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/hello-world/hello-world.ipynb) ![classification](https://user-images.githubusercontent.com/36741649/127172572-1cdab941-df5f-42e2-a367-2b334a3db6d8.jpg) diff --git a/notebooks/hugging-face-hub/README.md b/notebooks/hugging-face-hub/README.md index a100c7ababc..ac9e0fb95b1 100644 --- a/notebooks/hugging-face-hub/README.md +++ b/notebooks/hugging-face-hub/README.md @@ -1,6 +1,6 @@ # 🤗 Hugging Face Model Hub with OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fhugging-face-hub%2Fhugging-face-hub.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/hugging-face-hub/hugging-face-hub.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/hugging-face-hub/hugging-face-hub.ipynb) The Hugging Face (HF) Model Hub is a central repository for pre-trained deep learning models. It allows exploration and provides access to thousands of models for a wide range of tasks, including text classification, question answering, and image classification. Hugging Face provides Python packages that serve as APIs and tools to easily download and fine tune state-of-the-art pretrained models, namely [transformers] and [diffusers] packages. diff --git a/notebooks/image-classification-quantization/README.md b/notebooks/image-classification-quantization/README.md index 49bb3905e2b..396aab1aae0 100644 --- a/notebooks/image-classification-quantization/README.md +++ b/notebooks/image-classification-quantization/README.md @@ -1,7 +1,7 @@ # Accelerate Inference of MobileNet V2 Image Classification Model with NNCF in OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/openvinotoolkit/openvino_notebooks/HEAD?labpath=notebooks%2Fimage-classification-quantization%2Fimage-classification-quantization.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/image-classification-quantization/image-classification-quantization.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/image-classification-quantization/image-classification-quantization.ipynb) This tutorial demonstrates how to apply `INT8` quantization to the MobileNet V2 Image Classification model, using the [NNCF Post-Training Quantization API](https://docs.openvino.ai/2024/openvino-workflow/model-optimization-guide/quantizing-models-post-training.html). The tutorial uses [MobileNetV2](https://pytorch.org/vision/stable/_modules/torchvision/models/mobilenetv2.html) and [Cifar10 dataset](https://www.cs.toronto.edu/~kriz/cifar.html). diff --git a/notebooks/instruct-pix2pix-image-editing/instruct-pix2pix-image-editing.ipynb b/notebooks/instruct-pix2pix-image-editing/instruct-pix2pix-image-editing.ipynb index d4d016c9960..78b8d49bc60 100644 --- a/notebooks/instruct-pix2pix-image-editing/instruct-pix2pix-image-editing.ipynb +++ b/notebooks/instruct-pix2pix-image-editing/instruct-pix2pix-image-editing.ipynb @@ -1622,7 +1622,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/instruct-pix2pix-image-editing/instruct-pix2pix-image-editing.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/instruct-pix2pix-image-editing/instruct-pix2pix-image-editing.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/knowledge-graphs-conve/README.md b/notebooks/knowledge-graphs-conve/README.md index b12aa4aeb23..563da0e38a8 100644 --- a/notebooks/knowledge-graphs-conve/README.md +++ b/notebooks/knowledge-graphs-conve/README.md @@ -1,7 +1,7 @@ # Knowledge graphs model optimization using the Intel OpenVINO toolkit [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/openvinotoolkit/openvino_notebooks/HEAD?labpath=notebooks%2Fknowledge-graphs-conve%2Fknowledge-graphs-conve.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/knowledge-graphs-conve/knowledge-graphs-conve.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/knowledge-graphs-conve/knowledge-graphs-conve.ipynb) [Knowledge graphs](https://arxiv.org/pdf/2002.00388.pdf) provide an efficient way to represent domain information and are used in a number of real-world applications.

diff --git a/notebooks/language-quantize-bert/README.md b/notebooks/language-quantize-bert/README.md index 68615bf3451..3ae1443ce07 100644 --- a/notebooks/language-quantize-bert/README.md +++ b/notebooks/language-quantize-bert/README.md @@ -1,6 +1,6 @@ # Accelerate Inference of NLP models with Post-Training Quantization API of NNCF -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/language-quantize-bert/language-quantize-bert.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/language-quantize-bert/language-quantize-bert.ipynb) This tutorial demonstrates how to apply INT8 quantization to the Natural Language Processing model BERT, using the [Post-Training Quantization API](https://docs.openvino.ai/2024/openvino-workflow/model-optimization-guide/quantizing-models-post-training/basic-quantization-flow.html). diff --git a/notebooks/latent-consistency-models-image-generation/latent-consistency-models-image-generation.ipynb b/notebooks/latent-consistency-models-image-generation/latent-consistency-models-image-generation.ipynb index 67c8ee4f09b..5fbfd03aafa 100644 --- a/notebooks/latent-consistency-models-image-generation/latent-consistency-models-image-generation.ipynb +++ b/notebooks/latent-consistency-models-image-generation/latent-consistency-models-image-generation.ipynb @@ -1691,7 +1691,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/latent-consistency-models-image-generation/latent-consistency-models-image-generation.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/latent-consistency-models-image-generation/latent-consistency-models-image-generation.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/latent-consistency-models-image-generation/latent-consistency-models-optimum-demo.ipynb b/notebooks/latent-consistency-models-image-generation/latent-consistency-models-optimum-demo.ipynb index 718ba1a3e43..44d84de665a 100644 --- a/notebooks/latent-consistency-models-image-generation/latent-consistency-models-optimum-demo.ipynb +++ b/notebooks/latent-consistency-models-image-generation/latent-consistency-models-optimum-demo.ipynb @@ -359,7 +359,7 @@ "version": "3.11.7" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/llava-multimodal-chatbot/llava-multimodal-chatbot.ipynb b/notebooks/llava-multimodal-chatbot/llava-multimodal-chatbot.ipynb index 579388756e6..48a77df75e9 100644 --- a/notebooks/llava-multimodal-chatbot/llava-multimodal-chatbot.ipynb +++ b/notebooks/llava-multimodal-chatbot/llava-multimodal-chatbot.ipynb @@ -1563,7 +1563,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/llava-multimodal-chatbot/llava-multimodal-chatbot.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/llava-multimodal-chatbot/llava-multimodal-chatbot.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/machine-translation/README.md b/notebooks/machine-translation/README.md index 260a9f02be8..e5cfefe2cf2 100644 --- a/notebooks/machine-translation/README.md +++ b/notebooks/machine-translation/README.md @@ -1,6 +1,6 @@ # Interactive Machine Translation with OpenVINO [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/openvinotoolkit/openvino_notebooks/HEAD?labpath=notebooks%2Fmachine-translation%2Fmachine-translation.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/machine-translation/machine-translation.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/machine-translation/machine-translation.ipynb) Machine translation is a form of natural language processing that translates sentences from one language to another. diff --git a/notebooks/magika-content-type-recognition/README.md b/notebooks/magika-content-type-recognition/README.md index 84b2014d76c..e0699e7bf0e 100644 --- a/notebooks/magika-content-type-recognition/README.md +++ b/notebooks/magika-content-type-recognition/README.md @@ -1,7 +1,7 @@ # Magika: AI powered fast and efficient file type identification using OpenVINO [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fmagika-content-type-recognition%2Fmagika-content-type-recognition.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/magika-content-type-recognition/magika-content-type-recognition.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/magika-content-type-recognition/magika-content-type-recognition.ipynb) Magika is a novel AI powered file type detection tool that relies on the recent advance of deep learning to provide accurate detection. Under the hood, Magika employs a custom, highly optimized model that only weighs about 1MB, and enables precise file identification within milliseconds, even when running on a single CPU. diff --git a/notebooks/mobileclip-video-search/README.md b/notebooks/mobileclip-video-search/README.md index 4bec988f1e3..453581cfc96 100644 --- a/notebooks/mobileclip-video-search/README.md +++ b/notebooks/mobileclip-video-search/README.md @@ -1,5 +1,5 @@ # Visual Content Search using MobileCLIP and OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/mobileclip-video-search/mobileclip-video-search.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/mobileclip-video-search/mobileclip-video-search.ipynb) ![example.png](https://github.com/openvinotoolkit/openvino_notebooks/assets/29454499/4e241f82-548e-41c2-b1f4-45b319d3e519) diff --git a/notebooks/model-tools/README.md b/notebooks/model-tools/README.md index dc92ec8e2db..b9949470535 100644 --- a/notebooks/model-tools/README.md +++ b/notebooks/model-tools/README.md @@ -1,7 +1,7 @@ # Open Model Zoo Tools Tutorial [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fmodel-tools%2Fmodel-tools.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/model-tools/model-tools.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/model-tools/model-tools.ipynb) This notebook demonstrates how to download a model from Open Model Zoo, convert it to OpenVINO™ IR format, show information about the model, and benchmark the model. diff --git a/notebooks/music-generation/README.md b/notebooks/music-generation/README.md index 7b4b3b92b56..6aeb11b8a55 100644 --- a/notebooks/music-generation/README.md +++ b/notebooks/music-generation/README.md @@ -1,7 +1,7 @@ # Controllable Music Generation with MusicGen and OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fmusic-generation%2Fmusic-generation.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/music-generation/music-generation.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/music-generation/music-generation.ipynb) MusicGen is a single-stage auto-regressive Transformer model capable of generating high-quality music samples conditioned on text descriptions or audio prompts. The text prompt is passed to a text encoder model (T5) to obtain a sequence of hidden-state representations. These hidden states are fed to MusicGen, which predicts discrete audio tokens (audio codes). Finally, audio tokens are then decoded using an audio compression model (EnCodec) to recover the audio waveform. diff --git a/notebooks/music-generation/music-generation.ipynb b/notebooks/music-generation/music-generation.ipynb index 9b7b64a431b..adb36e669c7 100644 --- a/notebooks/music-generation/music-generation.ipynb +++ b/notebooks/music-generation/music-generation.ipynb @@ -790,7 +790,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/music-generation/music-generation.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/music-generation/music-generation.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/named-entity-recognition/README.md b/notebooks/named-entity-recognition/README.md index ec87bd59769..af6741cfc52 100644 --- a/notebooks/named-entity-recognition/README.md +++ b/notebooks/named-entity-recognition/README.md @@ -1,5 +1,5 @@ # Named entity recognition with OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/named-entity-recognition/named-entity-recognition.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/named-entity-recognition/named-entity-recognition.ipynb) Named entity recognition (NER) is one of the most popular data processing tasks. It is a natural language processing (NLP) method that involves detecting key information in unstructured text and categorizing it into pre-defined categories. These categories or named entities refer to the key subjects of the text, such as names, locations, companies and etc. NER is a suitable method for situations when a high-level overview of a large amount of text is needed. NER can be helpful with such tasks as analyzing key information in unstructured text or automating the information extraction of large amounts of data. diff --git a/notebooks/object-detection-webcam/README.md b/notebooks/object-detection-webcam/README.md index f1e7c14e95f..7927d99b757 100644 --- a/notebooks/object-detection-webcam/README.md +++ b/notebooks/object-detection-webcam/README.md @@ -1,7 +1,7 @@ # Live Object Detection with OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fobject-detection-webcam%2Fobject-detection.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/object-detection-webcam/object-detection.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/object-detection-webcam/object-detection.ipynb) *Binder is a free service where the webcam will not work, and performance on the video will not be good. For the best performance, install the notebooks locally.* diff --git a/notebooks/oneformer-segmentation/oneformer-segmentation.ipynb b/notebooks/oneformer-segmentation/oneformer-segmentation.ipynb index de203b9fa4e..1fa7460c49d 100644 --- a/notebooks/oneformer-segmentation/oneformer-segmentation.ipynb +++ b/notebooks/oneformer-segmentation/oneformer-segmentation.ipynb @@ -1304,7 +1304,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/oneformer-segmentation/oneformer-segmentation.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/oneformer-segmentation/oneformer-segmentation.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/openvino-api/README.md b/notebooks/openvino-api/README.md index 467dd20ac35..780252ea4b9 100644 --- a/notebooks/openvino-api/README.md +++ b/notebooks/openvino-api/README.md @@ -1,7 +1,7 @@ # OpenVINO™ API tutorial [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fopenvino-api%2Fopenvino-api.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/openvino-api/openvino-api.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/openvino-api/openvino-api.ipynb) This notebook explains the basics of the OpenVINO Runtime API. diff --git a/notebooks/openvino-tokenizers/README.md b/notebooks/openvino-tokenizers/README.md index 6d5e1d10cc3..907a7cd42fa 100644 --- a/notebooks/openvino-tokenizers/README.md +++ b/notebooks/openvino-tokenizers/README.md @@ -1,7 +1,7 @@ # OpenVINO Tokenizers: Incorporate Text Processing Into OpenVINO Pipelines [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fopenvino-tokenizers%2Fopenvino-tokenizers.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/openvino-tokenizers/openvino-tokenizers.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/openvino-tokenizers/openvino-tokenizers.ipynb)

diff --git a/notebooks/openvoice/README.md b/notebooks/openvoice/README.md index 9702546b02d..efe96d815ad 100644 --- a/notebooks/openvoice/README.md +++ b/notebooks/openvoice/README.md @@ -1,7 +1,7 @@ # Voice tone cloning with OpenVoice and OpenVINO [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fopenvoice%2Fopenvoice.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/openvoice/openvoice.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/openvoice/openvoice.ipynb) ![sdf](https://github.com/openvinotoolkit/openvino_notebooks/assets/5703039/ca7eab80-148d-45b0-84e8-a5a279846b51) diff --git a/notebooks/optical-character-recognition/README.md b/notebooks/optical-character-recognition/README.md index c87f89178e2..80e29f62107 100644 --- a/notebooks/optical-character-recognition/README.md +++ b/notebooks/optical-character-recognition/README.md @@ -1,6 +1,6 @@ # Optical Character Recognition (OCR) with OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/optical-character-recognition/optical-character-recognition.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/optical-character-recognition/optical-character-recognition.ipynb) | | | |---|---| diff --git a/notebooks/optimize-preprocessing/README.md b/notebooks/optimize-preprocessing/README.md index c54c36ac4eb..25edf03ff38 100644 --- a/notebooks/optimize-preprocessing/README.md +++ b/notebooks/optimize-preprocessing/README.md @@ -1,6 +1,6 @@ # Optimize preprocessing of image for the googlenet-v2 Image Classification Model with Preprocessing API in OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/optimize-preprocessing/optimize-preprocessing.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/optimize-preprocessing/optimize-preprocessing.ipynb) This tutorial demonstrates how the image could be transform to the data format expected by the model with Preprocessing API. Preprocessing API is an easy-to-use instrument, that enables integration of preprocessing steps into an execution graph and perform it on selected device, which can improve of device utilization. For more information about Preprocessing API, please, see this [overview](https://docs.openvino.ai/2024/openvino-workflow/running-inference/optimize-inference/optimize-preprocessing.html#) and [details](https://docs.openvino.ai/2024/openvino-workflow/running-inference/optimize-inference/optimize-preprocessing/preprocessing-api-details.html). The tutorial uses [InceptionResNetV2](https://www.tensorflow.org/api_docs/python/tf/keras/applications/inception_resnet_v2) model. diff --git a/notebooks/paddle-ocr-webcam/README.md b/notebooks/paddle-ocr-webcam/README.md index 7a54baf15f1..1fbe4c79a64 100644 --- a/notebooks/paddle-ocr-webcam/README.md +++ b/notebooks/paddle-ocr-webcam/README.md @@ -1,6 +1,6 @@ # PaddleOCR with OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/openvinotoolkit/openvino_notebooks/HEAD?labpath=notebooks%2Fpaddle-ocr-webcam%2Fpaddle-ocr-webcam.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/paddle-ocr-webcam/paddle-ocr-webcam.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/paddle-ocr-webcam/paddle-ocr-webcam.ipynb)

diff --git a/notebooks/paddle-ocr-webcam/paddle-ocr-webcam.ipynb b/notebooks/paddle-ocr-webcam/paddle-ocr-webcam.ipynb index c3ab1df1144..e140d870bee 100644 --- a/notebooks/paddle-ocr-webcam/paddle-ocr-webcam.ipynb +++ b/notebooks/paddle-ocr-webcam/paddle-ocr-webcam.ipynb @@ -730,7 +730,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/paddle-ocr-webcam/paddle-ocr-webcam.gif?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/paddle-ocr-webcam/paddle-ocr-webcam.gif?raw=true", "tags": { "categories": [ "Live Demos" diff --git a/notebooks/paddle-to-openvino/README.md b/notebooks/paddle-to-openvino/README.md index bf763a4c4c7..fe1a3abf703 100644 --- a/notebooks/paddle-to-openvino/README.md +++ b/notebooks/paddle-to-openvino/README.md @@ -1,7 +1,7 @@ # PaddlePaddle to OpenVINO™ IR Tutorial [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fpaddle-to-openvino%2Fpaddle-to-openvino-classification.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/paddle-to-openvino/paddle-to-openvino-classification.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/paddle-to-openvino/paddle-to-openvino-classification.ipynb) ![PaddlePaddle Classification](https://user-images.githubusercontent.com/77325899/127503530-72c8ce57-ef6f-40a7-808a-d7bdef909d11.png) diff --git a/notebooks/paint-by-example/paint-by-example.ipynb b/notebooks/paint-by-example/paint-by-example.ipynb index 4c054bbd072..7af4fe689f3 100644 --- a/notebooks/paint-by-example/paint-by-example.ipynb +++ b/notebooks/paint-by-example/paint-by-example.ipynb @@ -1696,7 +1696,7 @@ "version": "3.10.12" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/paint-by-example/paint-by-example.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/paint-by-example/paint-by-example.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/performance-tricks/README.md b/notebooks/performance-tricks/README.md index 12a12b1c76d..d60f461bcd3 100644 --- a/notebooks/performance-tricks/README.md +++ b/notebooks/performance-tricks/README.md @@ -9,7 +9,7 @@ In this notebook, we’ll show a set of performance tricks for optimizing infere 1. [Performance tricks in OpenVINO for latency mode](latency-tricks.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/performance-tricks/latency-tricks.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/performance-tricks/latency-tricks.ipynb) This notebook demonstrates how to optimize the inference latency in OpenVINO™. A set of optimization tricks, including model conversion with different data precision, “AUTO” device with latency mode, shared memory, inference with a further configuration, inference on GPU, etc., are introduced. @@ -17,7 +17,7 @@ This notebook demonstrates how to optimize the inference latency in OpenVINO™. 1. [Performance tricks in OpenVINO for throughput mode](throughput-tricks.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/performance-tricks/throughput-tricks.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/performance-tricks/throughput-tricks.ipynb) This notebook demonstrates how to optimize the inference throughput in OpenVINO™. A set of optimization tricks, including bigger batch size, “AUTO” device with throughput and cumulative throughput mode, asynchronous inference mode, etc., are introduced. diff --git a/notebooks/person-tracking-webcam/README.md b/notebooks/person-tracking-webcam/README.md index e697f3d7638..434c0d7018a 100644 --- a/notebooks/person-tracking-webcam/README.md +++ b/notebooks/person-tracking-webcam/README.md @@ -1,7 +1,7 @@ # Live Person Tracking with OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fperson-tracking-webcam%2Fperson-tracking.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/person-tracking-webcam/person-tracking.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/person-tracking-webcam/person-tracking.ipynb) *Binder is a free service where the webcam will not work, and performance on the video will not be good. For the best performance, install the notebooks locally.* diff --git a/notebooks/person-tracking-webcam/person-tracking.ipynb b/notebooks/person-tracking-webcam/person-tracking.ipynb index be568e768f7..6ca6cf9c7cb 100644 --- a/notebooks/person-tracking-webcam/person-tracking.ipynb +++ b/notebooks/person-tracking-webcam/person-tracking.ipynb @@ -749,7 +749,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/person-tracking-webcam/person-tracking.gif?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/person-tracking-webcam/person-tracking.gif?raw=true", "tags": { "categories": [ "Live Demos" diff --git a/notebooks/pix2struct-docvqa/README.md b/notebooks/pix2struct-docvqa/README.md index 75f4eeeafac..7a85c9a3a73 100644 --- a/notebooks/pix2struct-docvqa/README.md +++ b/notebooks/pix2struct-docvqa/README.md @@ -1,5 +1,5 @@ # Document Visual Question Answering Using Pix2Struct and OpenVINO -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/pix2struct-docvqa/pix2struct-docvqa.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/pix2struct-docvqa/pix2struct-docvqa.ipynb) DocVQA (Document Visual Question Answering) is a research field in computer vision and natural language processing that focuses on developing algorithms to answer questions related to the content of a document represented in image format, like a scanned document, screenshots, or an image of a text document. Unlike other types of visual question answering, where the focus is on answering questions related to images or videos, DocVQA is focused on understanding and answering questions based on the text and layout of a document. The questions can be about any aspect of the document text. DocVQA requires understanding the document’s visual content and the ability to read and comprehend the text in it. diff --git a/notebooks/pyannote-speaker-diarization/pyannote-speaker-diarization.ipynb b/notebooks/pyannote-speaker-diarization/pyannote-speaker-diarization.ipynb index 0a67634bec9..b48aff2977f 100644 --- a/notebooks/pyannote-speaker-diarization/pyannote-speaker-diarization.ipynb +++ b/notebooks/pyannote-speaker-diarization/pyannote-speaker-diarization.ipynb @@ -600,7 +600,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/pyannote-speaker-diarization/pyannote-speaker-diarization.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/pyannote-speaker-diarization/pyannote-speaker-diarization.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/pytorch-to-openvino/README.md b/notebooks/pytorch-to-openvino/README.md index aca08c991b1..1bee2aafc0e 100644 --- a/notebooks/pytorch-to-openvino/README.md +++ b/notebooks/pytorch-to-openvino/README.md @@ -6,14 +6,14 @@ This tutorial demonstrates how to convert PyTorch models to OpenVINO Intermediat * [pytorch-to-openvino](./pytorch-to-openvino.ipynb) shows how to convert the Pytorch model in formats `torch.nn.Module` and `torch.jit.ScriptModule` into OpenVINO Intermediate Representation. The tutorial uses [RegNetY_800MF](https://arxiv.org/abs/2003.13678) model from [torchvision](https://pytorch.org/vision/stable/index.html) pre-trained on [ImageNet](https://www.image-net.org/) dataset to demonstrate how to convert PyTorch models to OpenVINO Intermediate Representation using Model Converter. It also shows how to do classification inference on an image, using [OpenVINO Runtime](https://docs.openvino.ai/2024/openvino-workflow/running-inference.html) and compares the results of the PyTorch model with the OpenVINO IR model. -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/pytorch-to-openvino/pytorch-to-openvino.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/pytorch-to-openvino/pytorch-to-openvino.ipynb) ![classification_result](https://user-images.githubusercontent.com/29454499/250586825-2a4a74a6-e091-4e47-8f29-59a72fe4975f.png) * [pytorch-onnx-to-openvino](./pytorch-onnx-to-openvino.ipynb) shows how to convert the PyTorch model to OpenVINO IR with the intermediate step of exporting PyTorch model to ONNX format. [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fpytorch-to-openvino%2Fpytorch-onnx-to-openvino.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/pytorch-to-openvino/pytorch-onnx-to-openvino.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/pytorch-to-openvino/pytorch-onnx-to-openvino.ipynb) ![segmentation result](https://user-images.githubusercontent.com/29454499/203723317-1716e3ca-b390-47e1-bb98-07b4d8d097a0.png) The notebook uses OpenVINO Model Converter (OVC) to convert the open-source Lite-RASPP semantic segmentation model with a MobileNet V3 Large backbone from [torchvision](https://pytorch.org/vision/main/models/lraspp.html), trained on [COCO](https://cocodataset.org) dataset images using 20 categories that are present in the [Pascal VOC](https://paperswithcode.com/dataset/pascal-voc) dataset, to OpenVINO IR. It also shows how to do segmentation inference on an image, using [OpenVINO Runtime](https://docs.openvino.ai/2024/openvino-workflow/running-inference.html) and compares the results of the PyTorch model with the OpenVINO IR model. diff --git a/notebooks/qrcode-monster/qrcode-monster.ipynb b/notebooks/qrcode-monster/qrcode-monster.ipynb index 5049d2ecc88..27dcfb2eaec 100644 --- a/notebooks/qrcode-monster/qrcode-monster.ipynb +++ b/notebooks/qrcode-monster/qrcode-monster.ipynb @@ -1610,7 +1610,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/qrcode-monster/qrcode-monster.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/qrcode-monster/qrcode-monster.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/question-answering/README.md b/notebooks/question-answering/README.md index ee32365f11d..920a1243b75 100644 --- a/notebooks/question-answering/README.md +++ b/notebooks/question-answering/README.md @@ -1,7 +1,7 @@ # Interactive Question Answering with OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fquestion-answering%2Fquestion-answering.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/question-answering/question-answering.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/question-answering/question-answering.ipynb) diff --git a/notebooks/riffusion-text-to-music/riffusion-text-to-music.ipynb b/notebooks/riffusion-text-to-music/riffusion-text-to-music.ipynb index b6fe248d4e3..550a463a887 100644 --- a/notebooks/riffusion-text-to-music/riffusion-text-to-music.ipynb +++ b/notebooks/riffusion-text-to-music/riffusion-text-to-music.ipynb @@ -680,7 +680,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/riffusion-text-to-music/riffusion-text-to-music.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/riffusion-text-to-music/riffusion-text-to-music.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/sdxl-turbo/sdxl-turbo.ipynb b/notebooks/sdxl-turbo/sdxl-turbo.ipynb index af8545b5190..52cd78d1605 100644 --- a/notebooks/sdxl-turbo/sdxl-turbo.ipynb +++ b/notebooks/sdxl-turbo/sdxl-turbo.ipynb @@ -1016,7 +1016,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/sdxl-turbo/sdxl-turbo.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/sdxl-turbo/sdxl-turbo.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/segmenter-semantic-segmentation/README.md b/notebooks/segmenter-semantic-segmentation/README.md index 88cf8751928..0f5e11e346a 100644 --- a/notebooks/segmenter-semantic-segmentation/README.md +++ b/notebooks/segmenter-semantic-segmentation/README.md @@ -1,6 +1,6 @@ # Semantic Segmentation with OpenVINO™ using Segmenter -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/segmenter-semantic-segmentation/segmenter-semantic-segmentation.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/segmenter-semantic-segmentation/segmenter-semantic-segmentation.ipynb) This notebook demonstrates how to convert and use [Segmenter](https://github.com/rstrudel/segmenter) PyTorch model with OpenVINO. diff --git a/notebooks/siglip-zero-shot-image-classification/README.md b/notebooks/siglip-zero-shot-image-classification/README.md index 0ab32cdbdd5..a3a01fe2402 100644 --- a/notebooks/siglip-zero-shot-image-classification/README.md +++ b/notebooks/siglip-zero-shot-image-classification/README.md @@ -1,6 +1,6 @@ # Zero-shot Image Classification with SigLIP -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/siglip-zero-shot-image-classification/siglip-zero-shot-image-classification.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/siglip-zero-shot-image-classification/siglip-zero-shot-image-classification.ipynb) Zero-shot image classification is a computer vision task with the goal to classify images into one of several classes without any prior training or knowledge of these classes. diff --git a/notebooks/siglip-zero-shot-image-classification/siglip-zero-shot-image-classification.ipynb b/notebooks/siglip-zero-shot-image-classification/siglip-zero-shot-image-classification.ipynb index 3b19bfedc84..cb9221ca570 100644 --- a/notebooks/siglip-zero-shot-image-classification/siglip-zero-shot-image-classification.ipynb +++ b/notebooks/siglip-zero-shot-image-classification/siglip-zero-shot-image-classification.ipynb @@ -7,7 +7,7 @@ "source": [ "# Zero-shot Image Classification with SigLIP\n", "\n", - "[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/siglip-zero-shot-image-classification/siglip-zero-shot-image-classification.ipynb)\n", + "[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/siglip-zero-shot-image-classification/siglip-zero-shot-image-classification.ipynb)\n", "\n", "Zero-shot image classification is a computer vision task to classify images into one of several classes without any prior training or knowledge of the classes.\n", "\n", diff --git a/notebooks/softvc-voice-conversion/README.md b/notebooks/softvc-voice-conversion/README.md index b703b578bea..293b7e20e74 100644 --- a/notebooks/softvc-voice-conversion/README.md +++ b/notebooks/softvc-voice-conversion/README.md @@ -1,6 +1,6 @@ # SoftVC VITS Singing Voice Conversion and OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/softvc-voice-conversion/softvc-voice-conversion.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/softvc-voice-conversion/softvc-voice-conversion.ipynb) This tutorial is based on SoftVC VITS Singing Voice Conversion project. The purpose of this project was to enable developers to have their beloved anime characters perform singing tasks. The developers' intention was to focus solely on fictional characters and avoid any involvement of real individuals, anything related to real individuals deviates from the developer's original intention. diff --git a/notebooks/sound-generation-audioldm2/sound-generation-audioldm2.ipynb b/notebooks/sound-generation-audioldm2/sound-generation-audioldm2.ipynb index 66185758502..716c3bfa5f7 100644 --- a/notebooks/sound-generation-audioldm2/sound-generation-audioldm2.ipynb +++ b/notebooks/sound-generation-audioldm2/sound-generation-audioldm2.ipynb @@ -918,7 +918,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/sound-generation-audioldm2/sound-generation-audioldm2.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/sound-generation-audioldm2/sound-generation-audioldm2.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/sparsity-optimization/README.md b/notebooks/sparsity-optimization/README.md index bdee8fae609..8de20809809 100644 --- a/notebooks/sparsity-optimization/README.md +++ b/notebooks/sparsity-optimization/README.md @@ -1,5 +1,5 @@ # Accelerate Inference of Sparse Transformer Models with OpenVINO™ and 4th Gen Intel® Xeon® Scalable Processors -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/sparsity-optimization/sparsity-optimization.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/sparsity-optimization/sparsity-optimization.ipynb) This tutorial demonstrates how to improve performance of sparse Transformer models with [OpenVINO](https://docs.openvino.ai/) on 4th Gen Intel® Xeon® Scalable processors. It uses a pre-trained model from the [Hugging Face Transformers](https://huggingface.co/docs/transformers/index) library and shows how to convert it to the OpenVINO™ IR format and run inference on a CPU, using a dedicated runtime option that enables sparsity optimizations. It also demonstrates how to get more performance stacking sparsity with 8-bit quantization. To simplify the user experience, the [Hugging Face Optimum](https://huggingface.co/docs/optimum) library is used to convert the model to the OpenVINO™ IR format and quantize it using [Neural Network Compression Framework](https://github.com/openvinotoolkit/nncf). diff --git a/notebooks/speculative-sampling/README.md b/notebooks/speculative-sampling/README.md index 36729363fd5..ac957c57f67 100644 --- a/notebooks/speculative-sampling/README.md +++ b/notebooks/speculative-sampling/README.md @@ -1,7 +1,7 @@ # Text Generation via Speculative Sampling, KV Caching, and OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/speculative-sampling/speculative-sampling.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/speculative-sampling/speculative-sampling.ipynb) As model sizes grow, Generative AI implementations require significant inference resources. This not only increases the cost per generation from a prompt, but also increases the power consumption used to serve such requests. diff --git a/notebooks/speech-recognition-quantization/README.md b/notebooks/speech-recognition-quantization/README.md index c3b008508b6..19b5e7b3f76 100644 --- a/notebooks/speech-recognition-quantization/README.md +++ b/notebooks/speech-recognition-quantization/README.md @@ -1,5 +1,5 @@ # Quantize Speech Recognition Models using NNCF PTQ API -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/speech-recognition-quantization/speech-recognition-quantization-data2vec.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/speech-recognition-quantization/speech-recognition-quantization-data2vec.ipynb) This tutorial demonstrates how to apply `INT8` quantization to the speech recognition models, using post-training quantization with [NNCF](https://docs.openvino.ai/2024/openvino-workflow/model-optimization-guide/quantizing-models-post-training.html) (Neural Network Compression Framework). diff --git a/notebooks/speech-to-text/README.md b/notebooks/speech-to-text/README.md index a034603cdce..155a3f6efe9 100644 --- a/notebooks/speech-to-text/README.md +++ b/notebooks/speech-to-text/README.md @@ -1,7 +1,7 @@ # Speech to Text with OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fspeech-to-text%2Fspeech-to-text.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/speech-to-text/speech-to-text.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/speech-to-text/speech-to-text.ipynb) ![output](https://user-images.githubusercontent.com/36741649/140987347-279de058-55d7-4772-b013-0f2b12deaa61.png) diff --git a/notebooks/stable-diffusion-ip-adapter/stable-diffusion-ip-adapter.ipynb b/notebooks/stable-diffusion-ip-adapter/stable-diffusion-ip-adapter.ipynb index 15406d5bfb0..ae5b7294496 100644 --- a/notebooks/stable-diffusion-ip-adapter/stable-diffusion-ip-adapter.ipynb +++ b/notebooks/stable-diffusion-ip-adapter/stable-diffusion-ip-adapter.ipynb @@ -1257,7 +1257,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-ip-adapter/stable-diffusion-ip-adapter.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-ip-adapter/stable-diffusion-ip-adapter.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/stable-diffusion-keras-cv/stable-diffusion-keras-cv.ipynb b/notebooks/stable-diffusion-keras-cv/stable-diffusion-keras-cv.ipynb index a36762c0fb6..c7ddb5492f4 100644 --- a/notebooks/stable-diffusion-keras-cv/stable-diffusion-keras-cv.ipynb +++ b/notebooks/stable-diffusion-keras-cv/stable-diffusion-keras-cv.ipynb @@ -8,7 +8,7 @@ "\n", "

Important note: This notebook requires python >= 3.9. Please make sure that your environment fulfill to this requirement before running it
\n", "\n", - "Stable Diffusion is a powerful, open-source text-to-image generation model. There are multiple implementations of this pipeline in different frameworks. Previously, we already considered how to convert and optimize [PyTorch Stable Diffusion using HuggingFace Diffusers library](https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-text-to-image/stable-diffusion-text-to-image.ipynb). In this tutorial, we consider how to convert and run [Stable Diffusion from KerasCV](https://www.tensorflow.org/tutorials/generative/generate_images_with_stable_diffusion) that employs graph mode execution, which enhances performance by leveraging graph optimization and enabling parallelism and in the same time maintains a user-friendly interface for image generation.\n", + "Stable Diffusion is a powerful, open-source text-to-image generation model. There are multiple implementations of this pipeline in different frameworks. Previously, we already considered how to convert and optimize [PyTorch Stable Diffusion using HuggingFace Diffusers library](https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-text-to-image/stable-diffusion-text-to-image.ipynb). In this tutorial, we consider how to convert and run [Stable Diffusion from KerasCV](https://www.tensorflow.org/tutorials/generative/generate_images_with_stable_diffusion) that employs graph mode execution, which enhances performance by leveraging graph optimization and enabling parallelism and in the same time maintains a user-friendly interface for image generation.\n", "\n", "#### Table of contents:\n", " - [Prerequisites](#Prerequisites)\n", diff --git a/notebooks/stable-diffusion-torchdynamo-backend/stable-diffusion-torchdynamo-backend.ipynb b/notebooks/stable-diffusion-torchdynamo-backend/stable-diffusion-torchdynamo-backend.ipynb index 75003b7630d..1e6f1c58ab8 100644 --- a/notebooks/stable-diffusion-torchdynamo-backend/stable-diffusion-torchdynamo-backend.ipynb +++ b/notebooks/stable-diffusion-torchdynamo-backend/stable-diffusion-torchdynamo-backend.ipynb @@ -507,7 +507,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-torchdynamo-backend/stable-diffusion-torchdynamo-backend.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-torchdynamo-backend/stable-diffusion-torchdynamo-backend.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/stable-diffusion-v2/stable-diffusion-v2-infinite-zoom.ipynb b/notebooks/stable-diffusion-v2/stable-diffusion-v2-infinite-zoom.ipynb index a6d84f6a965..193e40432e7 100644 --- a/notebooks/stable-diffusion-v2/stable-diffusion-v2-infinite-zoom.ipynb +++ b/notebooks/stable-diffusion-v2/stable-diffusion-v2-infinite-zoom.ipynb @@ -1349,7 +1349,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-v2/stable-diffusion-v2-infinite-zoom.gif?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-v2/stable-diffusion-v2-infinite-zoom.gif?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.ipynb b/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.ipynb index 9789f1f9d65..7e24414426f 100644 --- a/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.ipynb +++ b/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.ipynb @@ -253,7 +253,7 @@ "version": "3.10.11" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/stable-diffusion-v2/stable-diffusion-v2-text-to-image-demo.ipynb b/notebooks/stable-diffusion-v2/stable-diffusion-v2-text-to-image-demo.ipynb index 3621a6c5a65..5c17918a485 100644 --- a/notebooks/stable-diffusion-v2/stable-diffusion-v2-text-to-image-demo.ipynb +++ b/notebooks/stable-diffusion-v2/stable-diffusion-v2-text-to-image-demo.ipynb @@ -515,7 +515,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/stable-diffusion-v2/stable-diffusion-v2-text-to-image.ipynb b/notebooks/stable-diffusion-v2/stable-diffusion-v2-text-to-image.ipynb index c1832c363e5..845054d6eb7 100644 --- a/notebooks/stable-diffusion-v2/stable-diffusion-v2-text-to-image.ipynb +++ b/notebooks/stable-diffusion-v2/stable-diffusion-v2-text-to-image.ipynb @@ -1552,7 +1552,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-v2/stable-diffusion-v2-optimum-demo.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/stable-diffusion-xl/segmind-vegart.ipynb b/notebooks/stable-diffusion-xl/segmind-vegart.ipynb index 9fac2841791..145470f192c 100644 --- a/notebooks/stable-diffusion-xl/segmind-vegart.ipynb +++ b/notebooks/stable-diffusion-xl/segmind-vegart.ipynb @@ -927,7 +927,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-xl/stable-diffusion-xl.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-xl/stable-diffusion-xl.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/stable-diffusion-xl/ssd-b1.ipynb b/notebooks/stable-diffusion-xl/ssd-b1.ipynb index 24d3e8de5fb..6e6f403a6e0 100644 --- a/notebooks/stable-diffusion-xl/ssd-b1.ipynb +++ b/notebooks/stable-diffusion-xl/ssd-b1.ipynb @@ -1192,7 +1192,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-xl/stable-diffusion-xl.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-xl/stable-diffusion-xl.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/stable-diffusion-xl/stable-diffusion-xl.ipynb b/notebooks/stable-diffusion-xl/stable-diffusion-xl.ipynb index 924c32edd4d..4f387920c4a 100644 --- a/notebooks/stable-diffusion-xl/stable-diffusion-xl.ipynb +++ b/notebooks/stable-diffusion-xl/stable-diffusion-xl.ipynb @@ -877,7 +877,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/stable-diffusion-xl/stable-diffusion-xl.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-xl/stable-diffusion-xl.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/style-transfer-webcam/README.md b/notebooks/style-transfer-webcam/README.md index 40236ff13cf..8551e3e418e 100644 --- a/notebooks/style-transfer-webcam/README.md +++ b/notebooks/style-transfer-webcam/README.md @@ -1,6 +1,6 @@ # Live Style Transfer with OpenVINO™ [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fstyle-transfer-webcam%2Fstyle-transfer.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/style-transfer-webcam/style-transfer.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/style-transfer-webcam/style-transfer.ipynb) *Binder and Google Colab are a free services where the webcam will not work, and performance on the video will not be good. For the best performance run the notebook locally.* diff --git a/notebooks/style-transfer-webcam/style-transfer.ipynb b/notebooks/style-transfer-webcam/style-transfer.ipynb index c30fcca7900..9993217429b 100644 --- a/notebooks/style-transfer-webcam/style-transfer.ipynb +++ b/notebooks/style-transfer-webcam/style-transfer.ipynb @@ -537,7 +537,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/style-transfer-webcam/style-transfer.gif?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/style-transfer-webcam/style-transfer.gif?raw=true", "tags": { "categories": [ "Live Demos" diff --git a/notebooks/surya-line-level-text-detection/README.md b/notebooks/surya-line-level-text-detection/README.md index ede0834cdda..c1908d655b9 100644 --- a/notebooks/surya-line-level-text-detection/README.md +++ b/notebooks/surya-line-level-text-detection/README.md @@ -1,6 +1,6 @@ # Line-level text detection with Surya -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/surya-line-level-text-detection/surya-line-level-text-detection.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/surya-line-level-text-detection/surya-line-level-text-detection.ipynb) In this tutorial we will perform line-level text detection using [Surya](https://github.com/VikParuchuri/surya) toolkit and OpenVINO. diff --git a/notebooks/surya-line-level-text-detection/surya-line-level-text-detection.ipynb b/notebooks/surya-line-level-text-detection/surya-line-level-text-detection.ipynb index 257734d4752..e3621869133 100644 --- a/notebooks/surya-line-level-text-detection/surya-line-level-text-detection.ipynb +++ b/notebooks/surya-line-level-text-detection/surya-line-level-text-detection.ipynb @@ -9,7 +9,7 @@ "source": [ "# Line-level text detection with Surya\n", "\n", - "[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/surya-line-level-text-detection/surya-line-level-text-detection.ipynb)\n", + "[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/surya-line-level-text-detection/surya-line-level-text-detection.ipynb)\n", "\n", "
Important note: This notebook requires python >= 3.9. Please make sure that your environment fulfill to this requirement before running it
\n", "\n", diff --git a/notebooks/table-question-answering/README.md b/notebooks/table-question-answering/README.md index 37443a5cf90..e73f2e2de36 100644 --- a/notebooks/table-question-answering/README.md +++ b/notebooks/table-question-answering/README.md @@ -1,6 +1,6 @@ # Table Question Answering using TAPAS and OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/table-question-answering/table-question-answering.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/table-question-answering/table-question-answering.ipynb) Table Question Answering (Table QA) is the answering a question about an information on a given table. You can use the Table Question Answering models to simulate SQL execution by inputting a table. diff --git a/notebooks/tensorflow-classification-to-openvino/README.md b/notebooks/tensorflow-classification-to-openvino/README.md index 54df844b2ab..a1e99a64440 100644 --- a/notebooks/tensorflow-classification-to-openvino/README.md +++ b/notebooks/tensorflow-classification-to-openvino/README.md @@ -1,7 +1,7 @@ # TensorFlow to OpenVINO™ Model Conversion Tutorial [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Ftensorflow-classification-to-openvino%2Ftensorflow-classification-to-openvino.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/tensorflow-classification-to-openvino/tensorflow-classification-to-openvino.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/tensorflow-classification-to-openvino/tensorflow-classification-to-openvino.ipynb) diff --git a/notebooks/tensorflow-hub/README.md b/notebooks/tensorflow-hub/README.md index 5c9446ace7e..174bb5b5357 100644 --- a/notebooks/tensorflow-hub/README.md +++ b/notebooks/tensorflow-hub/README.md @@ -1,6 +1,6 @@ # TensorFlow Hub models + OpenVINO -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/tensorflow-hub/tensorflow-hub.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/tensorflow-hub/tensorflow-hub.ipynb) [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Ftensorflow-hub%2Ftensorflow-hub.ipynb) This tutorial demonstrates how to convert TensorFlow Hub models to OpenVINO Intermediate Representation. diff --git a/notebooks/tensorflow-hub/tensorflow-hub.ipynb b/notebooks/tensorflow-hub/tensorflow-hub.ipynb index 7d008bd3e1f..6d52ad8772f 100644 --- a/notebooks/tensorflow-hub/tensorflow-hub.ipynb +++ b/notebooks/tensorflow-hub/tensorflow-hub.ipynb @@ -6,7 +6,7 @@ "metadata": {}, "source": [ "# Convert of TensorFlow Hub models to OpenVINO Intermediate Representation (IR)\n", - "[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/tensorflow-hub/tensorflow-hub.ipynb) [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Ftensorflow-hub%2Ftensorflow-hub.ipynb)\n", + "[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/tensorflow-hub/tensorflow-hub.ipynb) [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Ftensorflow-hub%2Ftensorflow-hub.ipynb)\n", "\n", "This tutorial demonstrates step-by-step instructions on how to convert models loaded from TensorFlow Hub using OpenVINO Runtime.\n", "\n", diff --git a/notebooks/tensorflow-object-detection-to-openvino/README.md b/notebooks/tensorflow-object-detection-to-openvino/README.md index 78f8c3018fe..2a111145ba6 100644 --- a/notebooks/tensorflow-object-detection-to-openvino/README.md +++ b/notebooks/tensorflow-object-detection-to-openvino/README.md @@ -5,7 +5,7 @@ This tutorial explains how to convert [TensorFlow Object Detection](https://gith ## Object Detection [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Ftensorflow-object-detection-to-openvino%2Ftensorflow-object-detection-to-openvino.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/tensorflow-object-detection-to-openvino/tensorflow-object-detection-to-openvino.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/tensorflow-object-detection-to-openvino/tensorflow-object-detection-to-openvino.ipynb) @@ -23,7 +23,7 @@ For details, please refer to [Installation Guide](../../README.md). ## Instance segmentation [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Ftensorflow-object-detection-to-openvino%2Ftensorflow-instance-segmentation-to-openvino.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/tensorflow-object-detection-to-openvino/tensorflow-instance-segmentation-to-openvino.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/tensorflow-object-detection-to-openvino/tensorflow-instance-segmentation-to-openvino.ipynb) diff --git a/notebooks/tensorflow-quantization-aware-training/README.md b/notebooks/tensorflow-quantization-aware-training/README.md index 7b84c2d868a..d6cd0177a9d 100644 --- a/notebooks/tensorflow-quantization-aware-training/README.md +++ b/notebooks/tensorflow-quantization-aware-training/README.md @@ -1,5 +1,5 @@ # Optimizing TensorFlow models with Neural Network Compression Framework of OpenVINO™ by 8-bit quantization. -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/tensorflow-quantization-aware-training/tensorflow-quantization-aware-training.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/tensorflow-quantization-aware-training/tensorflow-quantization-aware-training.ipynb) This tutorial demonstrates how to use [NNCF](https://github.com/openvinotoolkit/nncf) 8-bit quantization to optimize the [TensorFlow](https://www.tensorflow.org) model for inference with [OpenVINO Toolkit](https://docs.openvino.ai/). diff --git a/notebooks/tensorflow-training-openvino/tensorflow-training-openvino-nncf.ipynb b/notebooks/tensorflow-training-openvino/tensorflow-training-openvino-nncf.ipynb index 0fd0262cdfe..9f5644037e2 100644 --- a/notebooks/tensorflow-training-openvino/tensorflow-training-openvino-nncf.ipynb +++ b/notebooks/tensorflow-training-openvino/tensorflow-training-openvino-nncf.ipynb @@ -637,7 +637,7 @@ "\n", "Benchmark App is a command line tool that measures raw inference performance for a specified OpenVINO IR model. Run `benchmark_app --help` to see a list of available parameters. By default, Benchmark App tests the performance of the model specified with the `-m` parameter with asynchronous inference on CPU, for one minute. Use the `-d` parameter to test performance on a different device, for example an Intel integrated Graphics (iGPU), and `-t` to set the number of seconds to run inference. See the [documentation](https://docs.openvino.ai/2024/learn-openvino/openvino-samples/benchmark-tool.html) for more information.\n", "\n", - "This tutorial uses a wrapper function from [Notebook Utils](https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/utils/notebook_utils.ipynb). It prints the `benchmark_app` command with the chosen parameters.\n", + "This tutorial uses a wrapper function from [Notebook Utils](https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/utils/notebook_utils.ipynb). It prints the `benchmark_app` command with the chosen parameters.\n", "\n", "In the next cells, inference speed will be measured for the original and quantized model on CPU. If an iGPU is available, inference speed will be measured for CPU+GPU as well. The number of seconds is set to 15.\n", "\n", diff --git a/notebooks/text-prediction/README.md b/notebooks/text-prediction/README.md index 3df3ff758f3..2ff173b3728 100644 --- a/notebooks/text-prediction/README.md +++ b/notebooks/text-prediction/README.md @@ -1,5 +1,5 @@ # Text Prediction with OpenVINO™ -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/text-prediction/text-prediction.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/text-prediction/text-prediction.ipynb) Text generation is a type of natural language processing that uses computational linguistics and artificial intelligence to automatically produce text that can meet specific communicative needs. diff --git a/notebooks/tflite-selfie-segmentation/README.md b/notebooks/tflite-selfie-segmentation/README.md index 2ed23fa9c04..872b0cdcb20 100644 --- a/notebooks/tflite-selfie-segmentation/README.md +++ b/notebooks/tflite-selfie-segmentation/README.md @@ -1,6 +1,6 @@ # Selfie Segmentation using TFLite and OpenVINO [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Ftflite-selfie-segmentation%2Ftflite-selfie-segmentation.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/tflite-selfie-segmentation/tflite-selfie-segmentation.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/tflite-selfie-segmentation/tflite-selfie-segmentation.ipynb) The Selfie segmentation pipeline allows developers to easily separate the background from users within a scene and focus on what matters. Adding cool effects to selfies or inserting your users into interesting background environments has never been easier. Besides photo editing, this technology is also important for video conferencing. It helps to blur or replace the background during video calls. diff --git a/notebooks/tflite-to-openvino/README.md b/notebooks/tflite-to-openvino/README.md index 6baafe07029..f82b29ce2ba 100644 --- a/notebooks/tflite-to-openvino/README.md +++ b/notebooks/tflite-to-openvino/README.md @@ -1,5 +1,5 @@ # TensorFlow Lite to OpenVINO™ Model Conversion Tutorial -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/tflite-to-openvino/tflite-to-openvino.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/tflite-to-openvino/tflite-to-openvino.ipynb) diff --git a/notebooks/tiny-sd-image-generation/README.md b/notebooks/tiny-sd-image-generation/README.md index f0aa34ead43..353cdbe20ed 100644 --- a/notebooks/tiny-sd-image-generation/README.md +++ b/notebooks/tiny-sd-image-generation/README.md @@ -1,5 +1,5 @@ # Image Generation with Tiny-SD -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/tiny-sd-image-generation/tiny-sd-image-generation.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/tiny-sd-image-generation/tiny-sd-image-generation.ipynb) In recent times, the AI community has witnessed a remarkable surge in the development of larger and more performant language models, such as Falcon 40B, LLaMa-2 70B, Falcon 40B, MPT 30B, and in the imaging domain with models like SD2.1 and SDXL. These advancements have undoubtedly pushed the boundaries of what AI can achieve, enabling highly versatile and state-of-the-art image generation and language understanding capabilities. However, the breakthrough of large models comes with substantial computational demands. To resolve this issue, recent research on efficient Stable Diffusion has prioritized reducing the number of sampling steps and utilizing network quantization. diff --git a/notebooks/tiny-sd-image-generation/tiny-sd-image-generation.ipynb b/notebooks/tiny-sd-image-generation/tiny-sd-image-generation.ipynb index 4cccda85d5a..dccc7ef982b 100644 --- a/notebooks/tiny-sd-image-generation/tiny-sd-image-generation.ipynb +++ b/notebooks/tiny-sd-image-generation/tiny-sd-image-generation.ipynb @@ -1431,7 +1431,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/tiny-sd-image-generation/tiny-sd-image-generation.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/tiny-sd-image-generation/tiny-sd-image-generation.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/typo-detector/typo-detector.ipynb b/notebooks/typo-detector/typo-detector.ipynb index e7233a4faad..92d4f0e3238 100644 --- a/notebooks/typo-detector/typo-detector.ipynb +++ b/notebooks/typo-detector/typo-detector.ipynb @@ -745,7 +745,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/typo-detector/typo-detector.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/typo-detector/typo-detector.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/vision-background-removal/README.md b/notebooks/vision-background-removal/README.md index 214ae37d82e..a1ead147081 100644 --- a/notebooks/vision-background-removal/README.md +++ b/notebooks/vision-background-removal/README.md @@ -1,7 +1,7 @@ # Background Removal Demo [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fvision-background-removal%2Fvision-background-removal.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/vision-background-removal/vision-background-removal.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/vision-background-removal/vision-background-removal.ipynb) This demo notebook shows image segmentation and removing/adding background with [U^2-Net](https://github.com/xuebinqin/U-2-Net) and OpenVINO™. diff --git a/notebooks/vision-background-removal/vision-background-removal.ipynb b/notebooks/vision-background-removal/vision-background-removal.ipynb index bc0f5438126..73f63d97ef6 100644 --- a/notebooks/vision-background-removal/vision-background-removal.ipynb +++ b/notebooks/vision-background-removal/vision-background-removal.ipynb @@ -708,7 +708,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/vision-background-removal/vision-background-removal.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/vision-background-removal/vision-background-removal.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/vision-monodepth/README.md b/notebooks/vision-monodepth/README.md index a2420d10efc..1c4906b1fee 100644 --- a/notebooks/vision-monodepth/README.md +++ b/notebooks/vision-monodepth/README.md @@ -1,7 +1,7 @@ # Monodepth Demo [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fvision-monodepth%2Fvision-monodepth.ipynb) -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/vision-monodepth/vision-monodepth.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/vision-monodepth/vision-monodepth.ipynb) ![monodepth](https://user-images.githubusercontent.com/36741649/127173017-a0bbcf75-db24-4d2c-81b9-616e04ab7cd9.gif) diff --git a/notebooks/vision-paddlegan-anime/README.md b/notebooks/vision-paddlegan-anime/README.md index 458275ec0cb..6f11c1325da 100644 --- a/notebooks/vision-paddlegan-anime/README.md +++ b/notebooks/vision-paddlegan-anime/README.md @@ -1,5 +1,5 @@ # Photo to Anime with PaddleGAN Tutorial -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/vision-paddlegan-anime/vision-paddlegan-anime.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/vision-paddlegan-anime/vision-paddlegan-anime.ipynb) ![anime coco](https://user-images.githubusercontent.com/15709723/123559130-04550100-d74f-11eb-819c-a02284654428.jpg) diff --git a/notebooks/vision-paddlegan-superresolution/README.md b/notebooks/vision-paddlegan-superresolution/README.md index 4400630ccac..e64dd8963e3 100644 --- a/notebooks/vision-paddlegan-superresolution/README.md +++ b/notebooks/vision-paddlegan-superresolution/README.md @@ -1,6 +1,6 @@ # Super Resolution with PaddleGAN and OpenVINO -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/vision-paddlegan-superresolution/vision-paddlegan-superresolution.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/vision-paddlegan-superresolution/vision-paddlegan-superresolution.ipynb) This notebook demonstrates converting the RealSR (real-world super-resolution) model from [PaddlePaddle/PaddleGAN](https://github.com/PaddlePaddle/PaddleGAN) to OpenVINO Intermediate Representation (OpenVINO IR) format, and shows inference results on both the PaddleGAN and OpenVINO IR models. diff --git a/notebooks/vision-paddlegan-superresolution/vision-paddlegan-superresolution.ipynb b/notebooks/vision-paddlegan-superresolution/vision-paddlegan-superresolution.ipynb index 1dcab1f0dc1..e25b102caaa 100644 --- a/notebooks/vision-paddlegan-superresolution/vision-paddlegan-superresolution.ipynb +++ b/notebooks/vision-paddlegan-superresolution/vision-paddlegan-superresolution.ipynb @@ -708,7 +708,7 @@ "version": "3.11.5" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/vision-superresolution/vision-superresolution-image.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/vision-superresolution/vision-superresolution-image.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/vision-superresolution/README.md b/notebooks/vision-superresolution/README.md index 92c1f419fcc..94ba01cff63 100644 --- a/notebooks/vision-superresolution/README.md +++ b/notebooks/vision-superresolution/README.md @@ -2,7 +2,7 @@ image | video | |---|---| -[![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fvision-superresolution%2Fvision-superresolution-image.ipynb)
[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/vision-superresolution/vision-superresolution-image.ipynb) | [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fvision-superresolution%2Fvision-superresolution-video.ipynb)
[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/vision-superresolution/vision-superresolution-video.ipynb) +[![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fvision-superresolution%2Fvision-superresolution-image.ipynb)
[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/vision-superresolution/vision-superresolution-image.ipynb) | [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/eaidova/openvino_notebooks_binder.git/main?urlpath=git-pull%3Frepo%3Dhttps%253A%252F%252Fgithub.com%252Fopenvinotoolkit%252Fopenvino_notebooks%26urlpath%3Dtree%252Fopenvino_notebooks%252Fnotebooks%2Fvision-superresolution%2Fvision-superresolution-video.ipynb)
[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/vision-superresolution/vision-superresolution-video.ipynb)

diff --git a/notebooks/vision-superresolution/vision-superresolution-image.ipynb b/notebooks/vision-superresolution/vision-superresolution-image.ipynb index 06f9ade6709..710fda0f2cc 100644 --- a/notebooks/vision-superresolution/vision-superresolution-image.ipynb +++ b/notebooks/vision-superresolution/vision-superresolution-image.ipynb @@ -935,7 +935,7 @@ "version": "3.11.5" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/vision-superresolution/vision-superresolution-image.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/vision-superresolution/vision-superresolution-image.png?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/vision-superresolution/vision-superresolution-video.ipynb b/notebooks/vision-superresolution/vision-superresolution-video.ipynb index 0153a62ccab..6cd73408923 100644 --- a/notebooks/vision-superresolution/vision-superresolution-video.ipynb +++ b/notebooks/vision-superresolution/vision-superresolution-video.ipynb @@ -617,7 +617,7 @@ "version": "3.11.5" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/vision-superresolution/vision-superresolution-video.gif?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/vision-superresolution/vision-superresolution-video.gif?raw=true", "tags": { "categories": [ "Model Demos" diff --git a/notebooks/whisper-subtitles-generation/README.md b/notebooks/whisper-subtitles-generation/README.md index b9dbc0d609f..1103b0e4537 100644 --- a/notebooks/whisper-subtitles-generation/README.md +++ b/notebooks/whisper-subtitles-generation/README.md @@ -1,5 +1,5 @@ # Video Subtitle Generation with OpenAI Whisper -[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/whisper-subtitles-generation/whisper-convert.ipynb) +[![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/whisper-subtitles-generation/whisper-convert.ipynb) [Whisper](https://openai.com/blog/whisper/) is a general-purpose speech recognition model from [OpenAI](https://openai.com/). The model is able to almost flawlessly transcribe speech across dozens of languages and even handle poor audio quality or excessive background noise. This notebook will run the model with OpenVINO to generate transcription of a video. diff --git a/notebooks/wuerstchen-image-generation/wuerstchen-image-generation.ipynb b/notebooks/wuerstchen-image-generation/wuerstchen-image-generation.ipynb index a5863d18976..d9aeaaf5634 100644 --- a/notebooks/wuerstchen-image-generation/wuerstchen-image-generation.ipynb +++ b/notebooks/wuerstchen-image-generation/wuerstchen-image-generation.ipynb @@ -1389,7 +1389,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/wuerstchen-image-generation/wuerstchen-image-generation.png?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/wuerstchen-image-generation/wuerstchen-image-generation.png?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/notebooks/yolov8-optimization/README.md b/notebooks/yolov8-optimization/README.md index 86a556ae40d..039b6467a2e 100644 --- a/notebooks/yolov8-optimization/README.md +++ b/notebooks/yolov8-optimization/README.md @@ -11,10 +11,10 @@ YOLO stands for “You Only Look Once”, it is a popular family of real-time ob This tutorial explains how to convert and optimize the [YOLOv8](https://github.com/ultralytics/) PyTorch models with OpenVINO. These tutorials are considered object detection, instance segmentation and keypoint detection scenarios: -- [yolov8-object-detection](./yolov8-object-detection.ipynb) [![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/yolov8-optimization/yolov8-object-detection.ipynb) -- [yolov8-instance-segmentation](./yolov8-instance-segmentation.ipynb) [![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/yolov8-optimization/yolov8-instance-segmentation.ipynb) -- [yolov8-keypoint-detection](./yolov8-keypoint-detection.ipynb) [![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/yolov8-optimization/yolov8-keypoint-detection.ipynb) -- [yolov8-obb](./yolov8-obb.ipynb) [![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/master/notebooks/yolov8-optimization/yolov8-obb.ipynb) +- [yolov8-object-detection](./yolov8-object-detection.ipynb) [![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/yolov8-optimization/yolov8-object-detection.ipynb) +- [yolov8-instance-segmentation](./yolov8-instance-segmentation.ipynb) [![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/yolov8-optimization/yolov8-instance-segmentation.ipynb) +- [yolov8-keypoint-detection](./yolov8-keypoint-detection.ipynb) [![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/yolov8-optimization/yolov8-keypoint-detection.ipynb) +- [yolov8-obb](./yolov8-obb.ipynb) [![Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/yolov8-optimization/yolov8-obb.ipynb) Each case tutorial consists of the following steps:: diff --git a/notebooks/zeroscope-text2video/zeroscope-text2video.ipynb b/notebooks/zeroscope-text2video/zeroscope-text2video.ipynb index 07860a1d17e..d4c093519e2 100644 --- a/notebooks/zeroscope-text2video/zeroscope-text2video.ipynb +++ b/notebooks/zeroscope-text2video/zeroscope-text2video.ipynb @@ -1206,7 +1206,7 @@ "version": "3.8.10" }, "openvino_notebooks": { - "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/zeroscope-text2video/zeroscope-text2video.gif?raw=true", + "imageUrl": "https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/zeroscope-text2video/zeroscope-text2video.gif?raw=true", "tags": { "categories": [ "Model Demos", diff --git a/selector/src/notebook-metadata/notebook-metadata-collector.js b/selector/src/notebook-metadata/notebook-metadata-collector.js index 2f4e027237a..bffac4fa1c7 100644 --- a/selector/src/notebook-metadata/notebook-metadata-collector.js +++ b/selector/src/notebook-metadata/notebook-metadata-collector.js @@ -58,7 +58,7 @@ export class NotebookMetadataCollector extends NotebookContentReader { * @returns {string} */ _getNotebookGitHubLink() { - return `https://github.com/openvinotoolkit/openvino_notebooks/blob/master/notebooks/${this._notebookFilePath}`; + return `https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/${this._notebookFilePath}`; } /**