You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Update tools for UIF 1.2
Update quickstart to wait for the server
Update readme links
Add shape to ImageInferenceRequest
Bump to 0.4.0
Bump up to ROCM 5.6.1
Exclude Py3.6 from wheels
Signed-off-by: Varun Sharma <[email protected]>
Copy file name to clipboardExpand all lines: README.rst
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -38,14 +38,14 @@ The AMD Inference Server is integrated with the following libraries out of the g
38
38
* TensorFlow and PyTorch models with `ZenDNN <https://developer.amd.com/zendnn/>`__ on CPUs (optimized for AMD CPUs)
39
39
* ONNX models with `MIGraphX <https://github.com/ROCmSoftwarePlatform/AMDMIGraphX>`__ on AMD GPUs
40
40
* XModel models with `Vitis AI <https://www.xilinx.com/products/design-tools/vitis/vitis-ai.html>`__ on AMD FPGAs
41
-
* A graph of computation including as pre- and post-processing can be written using `AKS <https://github.com/Xilinx/Vitis-AI/tree/v3.0/src/AKS>`__ on AMD FPGAs for end-to-end inference
41
+
* A graph of computation including as pre- and post-processing can be written using `AKS <https://github.com/Xilinx/Vitis-AI/tree/bbd45838d4a93f894cfc9f232140dc65af2398d1/src/AKS>`__ on AMD FPGAs for end-to-end inference
42
42
43
43
Quick Start Deployment and Inference
44
44
------------------------------------
45
45
46
46
The following example demonstrates how to deploy the server locally and run a sample inference.
47
47
This example runs on the CPU and does not require any special hardware.
48
-
You can see a more detailed version of this example in the `quickstart <https://xilinx.github.io/inference-server/main/quickstart_inference.html>`__.
48
+
You can see a more detailed version of this example in the `quickstart <https://xilinx.github.io/inference-server/main/quickstart.html>`__.
49
49
50
50
.. code-block:: bash
51
51
@@ -80,7 +80,7 @@ Learn more
80
80
81
81
The documentation for the AMD Inference Server is available `online <https://xilinx.github.io/inference-server/>`__.
82
82
83
-
Check out the quickstart guides online to help you get started based on your use case(s): `inference <https://xilinx.github.io/inference-server/main/quickstart_inference.html>`__, `deployment <https://xilinx.github.io/inference-server/main/quickstart_deployment.html>`__ and `development <https://xilinx.github.io/inference-server/main/quickstart_development.html>`__.
83
+
Check out the `quickstart <https://xilinx.github.io/inference-server/main/quickstart.html>`__ online to help you get started.
The endpoints for each model will be the name of the model in the ``config.toml``, which should match the name of the parent directory in the model repository.
187
194
In this example, it would be "resnet50".
@@ -195,7 +202,7 @@ Server deployment summary
195
202
After setting up the server as above, you have the following information:
196
203
197
204
* IP address: 127.0.0.1 since the server is running on the same machine where you will run the inference
198
-
* Ports: 8998 and 50051 for HTTP and gRPC, respectively. If you used ``--publish``, your port numbers may be different and you can see what they are using ``docker ps``.
205
+
* Ports: 8998 and 50051 for HTTP and gRPC, respectively. If you used ``--publish`` in the ``docker run`` command to remap the ports, your port numbers may be different and you can see what they are using ``docker ps``.
199
206
* Endpoint: "resnet50" since that is what the model name was used in the model repository and in the configuration file
200
207
201
208
The rest of this example will use these values in the sample code so substitute your own values if they are different.
@@ -239,21 +246,22 @@ These results are post-processed and the top 5 labels for the image are printed.
0 commit comments