Skip to content

Commit

Permalink
Deployed 6150576 to docs-develop with MkDocs 1.3.0 and mike 1.1.2
Browse files Browse the repository at this point in the history
  • Loading branch information
github-actions[bot] committed Feb 12, 2024
1 parent 18512b3 commit 738544e
Show file tree
Hide file tree
Showing 5 changed files with 153 additions and 147 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -3211,7 +3211,7 @@ <h2 id="configuring-the-container">Configuring the container</h2>
<p><strong>CPUs (optional)</strong> - This value specifies how many CPUs a container can use. Decimal values are allowed, so if set to 1.5, the container will use at most one and a half cpu resource.</p>
</li>
<li>
<p><strong>GPUs (optional)</strong> - This field configures how many Nvidia GPUs a container can use. Allowed values are 'all' or an integer number. If there's no Nvidia GPU installed, leave it empty. The Nvidia Container Toolkit must be installed on the system to correctly configure the service, otherwise the container will not start.</p>
<p><strong>GPUs (optional)</strong> - This field configures how many Nvidia GPUs a container can use. Allowed values are <code>all</code> or an integer number. If there's no Nvidia GPU installed, leave it empty. The Nvidia Container Toolkit must be installed on the system to correctly configure the service, otherwise the container will not start. If the Nvidia Container Runtime is used, leave the field empty.</p>
</li>
<li>
<p><strong>Volume Mount (optional)</strong> - This field accepts a comma-separated list of system-to-container file mounts. This allows for the container to access files on the host machine.</p>
Expand All @@ -3220,6 +3220,9 @@ <h2 id="configuring-the-container">Configuring the container</h2>
<p><strong>Peripheral Device (optional)</strong> - This field accepts a comma-separated list of device paths. This parameter allows devices to be passed through from the host to the container.</p>
</li>
<li>
<p><strong>Runtime (optional)</strong>: Specifies the fully qualified name of an alternate OCI-compatible runtime, which is used to run commands specified by the 'run' instruction. Example: <code>nvidia</code> corresponds to <code>--runtime=nvidia</code>. Note: when using the Nvidia Container Runtime, leave the <strong>GPUs</strong> field empty. The GPUs available on the system will be accessible from the container by default.</p>
</li>
<li>
<p><strong>Networking Mode (optional)</strong> - Use this field to specify what networking mode the container will use. Possible Drivers include: bridge, none, container:{container id}, host. Please note that this field is case-sensitive. This field can also be used to connect to any of the networks listed by the cli command <code>docker network ls</code>.</p>
</li>
<li>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3386,14 +3386,17 @@ <h2 id="triton-server-container-service-component">Triton Server Container Servi
<ul>
<li><strong>Container Image</strong>: The image the container will be created with.</li>
<li><strong>Container Image Tag</strong>: Describes which image version that should be used for creating the container.</li>
<li><strong>Nvidia Triton Server ports</strong>: the ports used to connect to the server for HTTP, GRPC, and Metrics services.</li>
<li><strong>Nvidia Triton Server ports</strong>: The ports used to connect to the server for HTTP, GRPC, and Metrics services.</li>
<li><strong>Local model repository path</strong>: Specify the path on the filesystem where the models are stored.</li>
<li><strong>Local model decryption password</strong>: Specify the password to be used for decrypting models stored in the model repository. If none is specified, models are supposed to be plaintext.</li>
<li><strong>Inference Models</strong>: a comma-separated list of inference model names that the server will load. The models have to be already present in the filesystem where the server is running. This option simply tells the server to load the given models from a local or remote repository.</li>
<li><strong>Inference Models</strong>: A comma-separated list of inference model names that the server will load. The models have to be already present in the filesystem where the server is running. This option simply tells the server to load the given models from a local or remote repository.</li>
<li><strong>Local Backends Path</strong>: Specifies the host filesystem path where the backends are stored. This folder will be mounted as a volume inside the Triton container and will override the existing backends. If left blank, the backends provided by the Triton container will be used.</li>
<li><strong>Optional configuration for the local backends</strong>: A semi-colon separated list of configuration for the backends. i.e. tensorflow,version=2;tensorflow,allow-soft-placement=false </li>
<li><strong>Memory</strong>: The maximum amount of memory the container can use in bytes. Set it as a positive integer, optionally followed by a suffix of b, k, m, g, to indicate bytes, kilobytes, megabytes, or gigabytes. The minimum allowed value is platform dependent (i.e. 6m). If left empty, the memory assigned to the container will be set to a default value by the native container orchestrator.</li>
<li><strong>CPUs</strong>: Specify how many CPUs the Triton container can use. Decimal values are allowed, so if set to 1.5, the container will use at most one and a half cpu resource.</li>
<li><strong>GPUs</strong>: Specify how many Nvidia GPUs the Triton container can use. Allowed values are 'all' or an integer number. If there's no Nvidia GPU installed, leave the field empty.</li>
<li><strong>GPUs</strong>: Specify how many Nvidia GPUs the Triton container can use. Allowed values are 'all' or an integer number. If there's no Nvidia GPU installed, leave the field empty. If the Nvidia Container Runtime is used, leave the field empty.</li>
<li><strong>Runtime</strong>: Specifies the fully qualified name of an alternate OCI-compatible runtime, which is used to run commands specified by the 'run' instruction for the Triton container. Example: <code>nvidia</code> corresponds to <code>--runtime=nvidia</code>. Note: when using the Nvidia Container Runtime, leave the <strong>GPUs</strong> field empty. The GPUs available on the system will be accessible from the container by default.</li>
<li><strong>Devices</strong>: A comma-separated list of device paths passed to the Triton server container (e.g. <code>/dev/video0</code>).</li>
<li><strong>Timeout (in seconds) for time consuming tasks</strong>: Timeout (in seconds) for time consuming tasks like server startup, shutdown or model load. If the task exceeds the timeout, the operation will be terminated with an error.</li>
<li><strong>Max. GRPC message size (bytes)</strong>: this field controls the maximum allowed size for the GRPC calls to the server instance.</li>
</ul>
Expand Down
2 changes: 1 addition & 1 deletion docs-develop/search/search_index.json

Large diffs are not rendered by default.

Loading

0 comments on commit 738544e

Please sign in to comment.