Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fixed all accessibility colors. Again. #23055

Merged
merged 47 commits into from
Dec 11, 2024
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
Show all changes
47 commits
Select commit Hold shift + click to select a range
8189132
Fixed all accessibility colors. Again.
MaanavD Dec 9, 2024
ed72c86
Removed ruby version req. for checklinks.
MaanavD Dec 9, 2024
376e895
Forced ruby version 3.3 (current latest).
MaanavD Dec 9, 2024
2a55fcf
fixed checklinks.
MaanavD Dec 9, 2024
c8f9c9b
Fixed ignore syntax.
MaanavD Dec 9, 2024
dbca8bd
Attempted fix checklinks.
MaanavD Dec 9, 2024
43c415d
Attempted fix checklinks again.
MaanavD Dec 9, 2024
6e2a114
checklinks fix v3.
MaanavD Dec 9, 2024
7ac7c06
Trying gh action for checklinks.
MaanavD Dec 9, 2024
1fe28d7
update ruby.
MaanavD Dec 9, 2024
32dbc95
Update htmlproofer.
MaanavD Dec 9, 2024
b71a8a4
node 22 for LTS.
MaanavD Dec 9, 2024
99ae397
Working! now to add the flags.
MaanavD Dec 9, 2024
bdd406f
trying to remove wrong flag.
MaanavD Dec 9, 2024
06de101
only check links.
MaanavD Dec 9, 2024
adef153
Updated checks syntax.
MaanavD Dec 9, 2024
251df49
Allow missing HREF.
MaanavD Dec 9, 2024
c433053
Don't check external hashes.
MaanavD Dec 9, 2024
79fdacc
replaced all instances of http with https
MaanavD Dec 9, 2024
4ee28c7
Fixed false.
MaanavD Dec 9, 2024
fcd0a48
Fixed false (again?)/
MaanavD Dec 9, 2024
d2f3328
removed false?
MaanavD Dec 9, 2024
cc02e5c
Trying again.
MaanavD Dec 9, 2024
0006f62
Trying again..
MaanavD Dec 9, 2024
95acbed
Trying again...
MaanavD Dec 9, 2024
e220f26
Ignore linkedin for spam.
MaanavD Dec 9, 2024
8b5e423
Trying different checklinks.
MaanavD Dec 10, 2024
cee8760
Trying different checklinks.
MaanavD Dec 10, 2024
8e2bbb8
Formatted using prettier.
MaanavD Dec 10, 2024
4d6646f
Using older htmlproofer.
MaanavD Dec 10, 2024
0f63d84
Using oldest htmlproofer.
MaanavD Dec 10, 2024
5d7f3b2
Tried older ruby.
MaanavD Dec 10, 2024
d9d0b86
attempting htmlproofer again.
MaanavD Dec 10, 2024
921db9f
Trying to fix formatting.
MaanavD Dec 10, 2024
cfcaf65
Trying again.
MaanavD Dec 10, 2024
73e2ee7
Trying again....
MaanavD Dec 10, 2024
b1d4539
Fix checks.
MaanavD Dec 10, 2024
8a3b7f1
doublequote links.
MaanavD Dec 10, 2024
2adc617
remove links only check.
MaanavD Dec 10, 2024
3b8da3e
don't check external hash.
MaanavD Dec 10, 2024
827f066
don't check external hash..
MaanavD Dec 10, 2024
68eca83
disable external hash check.
MaanavD Dec 10, 2024
650ca6a
no check external hash.
MaanavD Dec 10, 2024
0b109b5
block linkedin, allow_missing_href
MaanavD Dec 10, 2024
0ddcb8d
block linkedin properly.
MaanavD Dec 10, 2024
f9cbd18
fixed links.
MaanavD Dec 10, 2024
824c734
Fixed all links.
MaanavD Dec 10, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 4 additions & 3 deletions _sass/color_schemes/onnxruntime.scss
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,7 @@ $btn-primary-color: #226aca;
// }
// 2024 December Accessibility changes
.highlight .s { color: #3c7a3b ;}
.highlight .py {color: #a25f00;}
// Initial Theme
.highlight .hll { background-color: #ffffcc; }
.highlight { background: #ffffff; }
Expand All @@ -22,7 +23,7 @@ $btn-primary-color: #226aca;
.highlight .o { color: #333333; }
.highlight .ch { color: #707070 ; }
.highlight .cm { color: #707070 ; }
.highlight .cp { color: #557799; }
.highlight .cp { color: #507191; }
.highlight .cpf { color: #707070 ; }
.highlight .c1 { color: #707070 ; }
.highlight .cs { color: #cc0000; font-weight: bold; }
Expand Down Expand Up @@ -52,7 +53,7 @@ $btn-primary-color: #226aca;
.highlight .ni { color: #880000; font-weight: bold; }
.highlight .ne { font-weight: bold; color: #eb0000; }
.highlight .nf { color: #0066BB; font-weight: bold; }
.highlight .nl { font-weight: bold; color: #8f6f00; }
.highlight .nl { font-weight: bold; color: #876900; }
.highlight .nn { font-weight: bold; color: #0d77a2 ; }
.highlight .nt { color: #007700; }
.highlight .nv { color: #996633; }
Expand All @@ -68,7 +69,7 @@ $btn-primary-color: #226aca;
.highlight .sc { color: #0044DD; }
.highlight .dl { background-color: #fff0f0; }
.highlight .sd { color: #d54220; }
.highlight .s2 { background-color: #fff0f0; }
.highlight .s2 { color: #3c7a3b ; background-color: #fff0f0; }
.highlight .se { color: #666666; font-weight: bold; background-color: #fff0f0; }
.highlight .sh { background-color: #fff0f0; }
.highlight .si { background-color: #eeeeee; }
Expand Down
2 changes: 1 addition & 1 deletion package.json
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@
"prettier-plugin-svelte": "^2.10.1",
"svelte": "^4.0.5",
"svelte-check": "^3.4.3",
"svelte-highlight": "^7.3.0",
"svelte-highlight": "^7.7.0",
"svelte-preprocess-import-assets": "^1.0.1",
"svelte-sitemap": "^2.6.0",
"tailwindcss": "^3.3.3",
Expand Down
4 changes: 2 additions & 2 deletions src/routes/+layout.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -2,15 +2,15 @@
import '../app.css';
import Header from './components/header.svelte';
import Footer from './components/footer.svelte';
import oneLight from 'svelte-highlight/styles/one-light';
import a11yLight from 'svelte-highlight/styles/a11y-light';
import { fade } from 'svelte/transition';
import { page } from '$app/stores';
export let data;
const url = 'https://onnxruntime.ai';
</script>

<svelte:head>
{@html oneLight}
{@html a11yLight}
{#if !data.pathname.startsWith('/blogs/')}
<title
>ONNX Runtime | {data.pathname == '/'
Expand Down
40 changes: 20 additions & 20 deletions src/routes/blogs/accelerating-llama-2/+page.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -46,14 +46,14 @@
<div class="container mx-auto px-4 md:px-8 lg:px-48 pt-8">
<h1 class="text-5xl pb-2">Accelerating LLaMA-2 Inference with ONNX Runtime</h1>
<p class="text-neutral">
By: <a href="https://www.linkedin.com/in/kunal-v-16315b94" class="text-blue-700 underline"
By: <a href="https://www.linkedin.com/in/kunal-v-16315b94" class="text-blue-800 underline"
>Kunal Vaishnavi</a
>
and
<a href="https://www.linkedin.com/in/parinitaparinita/" class="text-blue-700 underline">Parinita Rahi</a>
<a href="https://www.linkedin.com/in/parinitaparinita/" class="text-blue-800 underline">Parinita Rahi</a>
</p>
<p class="text-neutral">
14TH NOVEMBER, 2023 <span class="italic text-stone-500">(Updated 22nd November)</span>
14TH NOVEMBER, 2023 <span class="italic text-stone-600">(Updated 22nd November)</span>
</p>
<div class="py-4">
<p class="mb-4">
Expand All @@ -71,13 +71,13 @@
quantization updates, and cross-platform usage scenarios.
</p>

<h2 class="text-blue-700 text-3xl mb-4">Background: Llama2 and Microsoft</h2>
<h2 class="text-blue-800 text-3xl mb-4">Background: Llama2 and Microsoft</h2>

<p class="mb-4">
Llama2 is a state-of-the-art open source LLM from Meta ranging in scale from 7B to 70B
parameters (7B, 13B, 70B). Microsoft and Meta <a
href="https://blogs.microsoft.com/blog/2023/07/18/microsoft-and-meta-expand-their-ai-partnership-with-llama-2-on-azure-and-windows/"
class="text-blue-700 underline">announced</a
class="text-blue-800 underline">announced</a
> their AI on Azure and Windows collaboration in July 2023. As part of the announcement, Llama2
was added to the Azure AI model catalog, which serves as a hub of foundation models that empower
developers and machine learning (ML) professionals to easily discover, evaluate, customize, and
Expand All @@ -90,7 +90,7 @@
your costs.
</p>

<h2 class="text-blue-700 text-3xl mb-4">
<h2 class="text-blue-800 text-3xl mb-4">
Faster Inferencing with New ONNX Runtime Optimizations
</h2>

Expand All @@ -116,7 +116,7 @@
</div>
<div class="mt-2 mb-4 text-center">Figure 1: E2E Throughput Comparisons</div>

<h2 class="text-blue-700 text-3xl mb-4">Latency and Throughput</h2>
<h2 class="text-blue-800 text-3xl mb-4">Latency and Throughput</h2>

<p class="mb-4">
The graphs below show latency comparisons between the ONNX Runtime and PyTorch variants of the
Expand Down Expand Up @@ -153,11 +153,11 @@
<p class="mb-4">
More details on these metrics can be found <a
href="https://github.com/microsoft/onnxruntime-inference-examples/blob/main/python/models/llama/README.md"
class="text-blue-700 underline">here</a
class="text-blue-800 underline">here</a
>.
</p>

<h2 class="text-blue-700 text-3xl mb-4">ONNX Runtime with Multi-GPU Inference</h2>
<h2 class="text-blue-800 text-3xl mb-4">ONNX Runtime with Multi-GPU Inference</h2>

<p class="mb-4">
ONNX Runtime supports multi-GPU inference to enable serving large models. Even in FP16
Expand All @@ -166,7 +166,7 @@
</p>

<p class="mb-4">
ONNX Runtime applied <a href="https://arxiv.org/pdf/1909.08053.pdf" class="text-blue-700 underline"
ONNX Runtime applied <a href="https://arxiv.org/pdf/1909.08053.pdf" class="text-blue-800 underline"
>Megatron-LM</a
>
Tensor Parallelism on the 70B model to split the original model weight onto different GPUs. Megatron
Expand All @@ -177,7 +177,7 @@
You can find additional example scripts
<a
href="https://github.com/microsoft/onnxruntime/blob/main/onnxruntime/python/tools/transformers/models/llama/"
class="text-blue-700 underline">here</a
class="text-blue-800 underline">here</a
>.
</p>

Expand All @@ -186,7 +186,7 @@
<figcaption class="mt-2 mb-4 text-center">Figure 4: 70B Llama2 Model Throughput</figcaption>
</figure>

<h2 class="text-blue-700 text-3xl mb-4">ONNX Runtime Optimizations</h2>
<h2 class="text-blue-800 text-3xl mb-4">ONNX Runtime Optimizations</h2>
<figure class="px-10 pt-4">
<img src={figure5} alt="LLaMA-2 Optimization Diagram" />
<figcaption class="mt-2 mb-4 text-center">Figure 5: LLaMA-2 Optimization Diagram</figcaption>
Expand Down Expand Up @@ -253,24 +253,24 @@
calculate the rotary embeddings more efficiently with less memory usage. The rotary embedding
compute kernels also support interleaved and non-interleaved formats to support both the <a
href="https://github.com/microsoft/Llama-2-Onnx"
class="text-blue-700 underline">Microsoft version of LLaMA-2</a
class="text-blue-800 underline">Microsoft version of LLaMA-2</a
>
and the Hugging Face version of LLaMA-2 respectively while sharing the same calculations.
</p>

<p class="mb-4">
The optimizations work for the <a
href="https://huggingface.co/meta-llama"
class="text-blue-700 underline">Hugging Face versions</a
class="text-blue-800 underline">Hugging Face versions</a
>
(models ending with <i>-hf</i>) and the Microsoft versions. You can download the optimized HF
versions from
<a href="https://github.com/microsoft/Llama-2-Onnx/tree/main-CUDA_CPU" class="text-blue-700 underline"
<a href="https://github.com/microsoft/Llama-2-Onnx/tree/main-CUDA_CPU" class="text-blue-800 underline"
>Microsoft's LLaMA-2 ONNX repository</a
>. Stay tuned for newer Microsoft versions coming soon!
</p>

<h2 class="text-blue-700 text-3xl mb-4">Optimize your own model using Olive</h2>
<h2 class="text-blue-800 text-3xl mb-4">Optimize your own model using Olive</h2>

<p class="mb-4">
Olive is a hardware-aware model optimization tool that incorporates advanced techniques such
Expand All @@ -282,25 +282,25 @@
<p class="mb-4">
Here is an example of <a
href="https://github.com/microsoft/Olive/tree/main/examples/llama2"
class="text-blue-700 underline">Llama2 optimization with Olive</a
class="text-blue-800 underline">Llama2 optimization with Olive</a
>, which harnesses ONNX Runtime optimizations highlighted in this blog. Distinct optimization
flows cater to various requirements. For instance, you have the flexibility to choose
different data types for quantization in CPU and GPU inference, based on your accuracy
tolerance. Additionally, you can fine-tune your own Llama2 model with Olive-QLoRa on client
GPUs and perform inference with ONNX Runtime optimizations.
</p>

<h2 class="text-blue-700 text-3xl mb-4">Usage Example</h2>
<h2 class="text-blue-800 text-3xl mb-4">Usage Example</h2>

<p class="mb-4">
Here is a <a
href="https://github.com/microsoft/onnxruntime-inference-examples/blob/main/python/models/llama/LLaMA-2%20E2E%20Notebook.ipynb"
class="text-blue-700 underline">sample notebook</a
class="text-blue-800 underline">sample notebook</a
> that shows you an end-to-end example of how you can use the above ONNX Runtime optimizations
in your application.
</p>

<h2 class="text-blue-700 text-3xl mb-4">Conclusion</h2>
<h2 class="text-blue-800 text-3xl mb-4">Conclusion</h2>

<p class="mb-4">
The advancements discussed in this blog provide faster Llama2 inferencing with ONNX Runtime,
Expand Down
2 changes: 1 addition & 1 deletion src/routes/blogs/blog-post-featured.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@
<h2 class="card-title">{title}</h2>
<p>{description}</p>
<img class="rounded" src={image} alt={imgalt} />
<div class="text-right text-blue-700">
<div class="text-right text-blue-800">
{date}
</div>
</div>
Expand Down
2 changes: 1 addition & 1 deletion src/routes/blogs/blog-post.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@
<div class="card-body">
<h2 class="card-title">{title}</h2>
<p>{description}</p>
<p class="text-blue-700 text-right">
<p class="text-blue-800 text-right">
{date}
</p>
</div>
Expand Down
4 changes: 2 additions & 2 deletions src/routes/blogs/post.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -82,15 +82,15 @@
<p class="inline">By:</p>
{/if}
{#each authors as author, i}
<a href={authorsLink[i]} class="text-blue-700">{author}</a>{i + 1 === authors.length
<a href={authorsLink[i]} class="text-blue-800">{author}</a>{i + 1 === authors.length
? ''
: ', '}
{/each}
</p>
<p class="text-neutral">
{date.toLocaleUpperCase()}
{#if updated != undefined}
<span class="italic text-stone-500">(Updated {updated})</span>
<span class="italic text-stone-600">(Updated {updated})</span>
{/if}
</p>
<div class="py-4 markdown-body">
Expand Down
4 changes: 0 additions & 4 deletions src/routes/components/code-blocks.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,6 @@
import FaLink from 'svelte-icons/fa/FaLink.svelte';
import { blur, fade } from 'svelte/transition';
import { d } from 'svelte-highlight/languages';
import github from "svelte-highlight/styles/github";

let pythonCode =
'import onnxruntime as ort\n# Load the model and create InferenceSession\nmodel_path = "path/to/your/onnx/model"\nsession = ort.InferenceSession(model_path)\n# "Load and preprocess the input image inputTensor"\n...\n# Run inference\noutputs = session.run(None, {"input": inputTensor})\nprint(outputs)';
Expand Down Expand Up @@ -51,9 +50,6 @@

</script>
<svelte:window bind:innerWidth/>
<svelte:head>
{@html github}
</svelte:head>
<div class="container mx-auto px-4">
<h3 class="text-xl mb-4 text-center">
Use ONNX Runtime with your favorite language and get started with the tutorials:
Expand Down
2 changes: 1 addition & 1 deletion src/routes/components/hero.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -126,7 +126,7 @@
<a class="underline" href="http://">More interested in training? More info here.</a>
</p> -->
<p class="text-lg mt-2">
<a class="text-blue-700 font-medium" href="./getting-started"
<a class="text-blue-800 font-medium" href="./getting-started"
>Interested in using other languages? See the many others we support →</a
>
</p>
Expand Down
2 changes: 1 addition & 1 deletion src/routes/components/performance.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
CPU, GPU, NPU - no matter what hardware you run on, ONNX Runtime optimizes for latency,
throughput, memory utilization, and binary size. In addition to excellent out-of-the-box
performance for common usage patterns, additional
<a href="https://onnxruntime.ai/docs/performance/" class="text-blue-700 underline"
<a href="https://onnxruntime.ai/docs/performance/" class="text-blue-800 underline"
>model optimization techniques</a
> and runtime configurations are available to further improve performance for specific use cases
and models.
Expand Down
4 changes: 2 additions & 2 deletions src/routes/components/training-and-inference.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -62,10 +62,10 @@
<p class="text-lg">
Accelerate training of popular models, including <a
href="https://huggingface.co/"
class="text-blue-700">Hugging Face</a
class="text-blue-800 underline">Hugging Face</a
>
models like Llama-2-7b and curated models from the
<a href="https://ml.azure.com/" class="text-blue-700"
<a href="https://ml.azure.com/" class="text-blue-800 underline"
>Azure AI | Machine Learning Studio</a
> model catalog.
</p>
Expand Down
18 changes: 9 additions & 9 deletions src/routes/components/winarm.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -35,33 +35,33 @@
<div class="divider" />
<div class="grid grid-cols-3 gap-4">
<div class="md:col-span-2 col-span-3">
<h2 class="text-xl text-blue-700">Get started on your Windows Dev Kit 2023 today</h2>
<h2 class="text-xl text-blue-800">Get started on your Windows Dev Kit 2023 today</h2>
Follow these steps to setup your device to use ONNX Runtime (ORT) with the built in NPU:
<ol class="list-decimal ml-10">
<li>
<a
class="text-blue-700"
class="text-blue-800"
href="https://qpm.qualcomm.com/main/tools/details/qualcomm_ai_engine_direct">Download</a
> the Qualcomm AI Engine Direct SDK (QNN SDK)
</li>
<li>
<a
class="text-blue-700"
class="text-blue-800"
href="https://www.nuget.org/packages/Microsoft.ML.OnnxRuntime.QNN">Download</a
> and install the ONNX Runtime with QNN package
</li>
<li>Start using the ONNX Runtime API in your application.</li>
</ol>
<br /><br />
<p class="text-xl text-blue-700">Optimizing models for the NPU</p>
<a class="text-blue-700" href="https://onnx.ai/">ONNX</a> is a standard format for
<p class="text-xl text-blue-800">Optimizing models for the NPU</p>
<a class="text-blue-800" href="https://onnx.ai/">ONNX</a> is a standard format for
representing ML models authored in frameworks like PyTorch, TensorFlow, and others. ONNX
Runtime can run any ONNX model, however to make use of the NPU, you currently need to quantize
the ONNX model to QDQ model.
<br />
See our
<a
class="text-blue-700"
class="text-blue-800"
href="https://github.com/microsoft/onnxruntime-inference-examples/tree/main/c_cxx/QNN_EP/mobilenetv2_classification"
>C# tutorial</a
>
Expand All @@ -70,13 +70,13 @@
Many models can be optimized for the NPU using this process. Even if a model cannot be optimized
for the NPU, it can still be run by ONNX Runtime on the CPU.
<br /><br />
<p class="text-xl text-blue-700">Getting Help</p>
<p class="text-xl text-blue-800">Getting Help</p>
For help with ONNX Runtime, you can<a
class="text-blue-700"
class="text-blue-800"
href="https://github.com/microsoft/onnxruntime/discussions">start a discussion</a
>
on GitHub or
<a class="text-blue-700" href="https://github.com/microsoft/onnxruntime/issues"
<a class="text-blue-800" href="https://github.com/microsoft/onnxruntime/issues"
>file an issue</a
>.
</div>
Expand Down
2 changes: 1 addition & 1 deletion src/routes/inference/+page.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -156,7 +156,7 @@
</div>
<a
href="https://github.com/microsoft/onnxruntime-inference-examples/tree/main/mobile"
class="text-2xl text-blue-700">See more examples of ONNX Runtime Mobile on GitHub. →</a
class="text-2xl text-blue-800">See more examples of ONNX Runtime Mobile on GitHub. →</a
>
</div>
</div>
Expand Down
2 changes: 1 addition & 1 deletion src/routes/models/+page.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -84,7 +84,7 @@
<h2 class="text-2xl">Can't find what you're looking for?</h2>
<p>
Convert to ONNX, optimize, and quantize your own models quickly and easily with <a
class="text-blue-700 underline"
class="text-blue-800 underline"
href="https://github.com/microsoft/Olive/tree/main">Olive</a
>. Here's a quick snippet showing you how easy it can be done:
</p>
Expand Down
2 changes: 1 addition & 1 deletion src/routes/onnx/+page.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
We hope your stay is short and that you quickly get what you need!
</p>
<p class="text-lg mb-4">
All below links are <a href="https://aka.ms" class="text-blue-700">aka.ms/</a> supported, so
All below links are <a href="https://aka.ms" class="text-blue-800">aka.ms/</a> supported, so
feel free to use those in the future.
</p>
</div>
Expand Down
Loading