Deep Learning Accelerator Programming Interface (nvm_dlaSample)
The NvMedia nvm_dlaSample
sample application demonstrates how to use the
NvMedia Deep Learning Accelerator (DLA) APIs to perform deep learning inference operations.
The sample uses the NVIDIA? SoC DLA hardware engine.
The nvm_dlaSample
application has four testing modes:
- Runtime mode.
- Single thread for running DLA.
- Demonstrates how to create and initialize DLA instances.
- Demonstrates how to run DLA with provided input data and network.
- SciSync mode.
- Single thread for running DLA. Two supporting threads for synchronization (signaler and waiter).
- Demonstrates how to create and initialize DLA instances.
- Demonstrates how to run DLA with provided input data and network.
- Demonstrates how to synchronize DLA task submission with a CPU signaler and CPU waiter.
- Multithreaded mode.
- Multiple threads (4) for running DLA.
- Demonstrates how to create and initialize DLA instances.
- Demonstrates how to run DLA with provided input data and network.
- Ping mode.
- Pings the specified instance if it exists.
Note: The tegrastats utility enables and reports on resource
utilization. For addition information, refer to DLA in GitHub.