Skip to content

Conversation

Deeksha-20-99
Copy link
Collaborator

@Deeksha-20-99 Deeksha-20-99 commented Sep 19, 2025

This demo showcases time-sensitive Lingua Franca demos that include two LLM agents and measure both logical and physical time per user query for LLM-based inference. In this demo, we also present the federated execution of time-sensitive LLM-based inference.

  • The Judge reactor takes in user input as a physical action, and this acts as a quiz question to the two LLM agents.
  • The LlmA and LlmB reactors receive the query from the Judge and each invokes its respective LLM model to generate an answer.
  • Both agents run concurrently. Whichever LLM responds first is declared the winner, and its answer is reported along with logical and physical execution times.

For this demo, the LLM models we have utilized are LLAMA chat models having a billion parameters, which are then quantized. These models are used for inference with the training knowledge, and no context is given.

We also explore federated execution, where each LLM agent could run as a separate federate, coordinated by the RTI. This follows the same principle of the game, except that we try to offload models on different devices (to support the models with higher parameters to be run on devices with greater GPU support).
This demo explores time-sensitive events and distributed execution of LLM-based systems using Lingua Franca.

llm/README.md Outdated

# Pre-requisites

You need Python installed, as llm.py is written in Python.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We need Python version information here (e.g., minimum version requirement).

@hokeun hokeun changed the title Demo LF program with latency sensitive LLM inferences Demo LF programs with latency sensitive LLM inferences Oct 16, 2025
@hokeun hokeun marked this pull request as ready for review October 17, 2025 19:17
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants