Hi Nitin,
Individual calls to the prompt / promptStreaming are queued as most devices are not capable of handling parallel LLM workloads.
An alternative would be to engineer the system prompt and user input so that it can process multiple inputs in one call, as long as everything fits within the context window.
As for function calling, it's not yet supported in the API and the current Nano model hasn't been specifically trained for that use case.
Anecdotal feedback suggests that it's viable enough for prototyping / exploring ideas.
Best,