Basics
Getting started
Installing the client is as simple as:
If your machine can run Python3 — it can run Everinfer client.
You will need a personal API key to use Everinfer. Reach out to us to get a demo key with some free compute power attached: hello@everinfer.ai We are very responsive, don't hesitate to do it regardless of your use case :)
Define your API key and an ONNX model that you'd like to use:
Hosting models
Import Everinfer Client
to manage your pipelines and create inference engines. Authenticate with your API key:
The next step is to upload a model, assign it a name, and, optionally, define some metadata JSON-serializable dictionary:
Then, create an inference engine:
That is it! Now you are ready to run inference on remote GPUs.
Running inference
engine
accepts a list of Python dicts in the following format:
Where 'input_name'
has to match the input name defined for your ONNX graph, and input can be a numpy
array or a path to .jpg
or .png
image.
Types of input have to match types expected by ONNX graph. For example, ONNX files that are exported with torch.FloatTensor()
as input expect np.float32
inputs.
Call the model:
Now you have a general understanding of Everinfer flow and its simplicity.
Take a look at the next example, which covers everything you need to fire up your own, production-ready pipeline, including chaining ONNX graphs, optimizing pre-processing, and more.
Last updated