=================== Neuron Run-Time API =================== .. contents:: Sections :local: :depth: 2 Neuron Runtime API provides a set of API functions that users can invoke from within a C/C++ program to create a run-time environment, parse a compiled model file, and perform on-device network inference. NeuroPilot 6 includes two versions of Neuron Runtime API. - :doc:`Neuron Runtime API V1 ` - :doc:`Neuron Runtime API V2 ` Runtime API Versions -------------------- **Neuron Runtime V1** * For single task and sequential execution (synchronous inference). * Inference API function: ``NeuronRuntime_inference`` (synchronous function call). * Use Neuron Runtime V1 if there is no time overlap between each inference. **Neuron Runtime V2** * For multi-task execution in parallel (asynchronous inference). * Inference API function: ``NeuronRuntimeV2_enqueue`` (asynchronous function call). * Use Neuron Runtime V2 if the next inference will start before the previous inference has finished. * Runtime V2 might increase power consumption, because parallel execution uses more hardware resources. * Runtime V2 might increase memory footprint, because each parallel task maintains its own working buffer.