Core module for Neon LLM's
API requests should include history
, a list of tuples of strings, and the current
query
Example Request:
{ "history": [["user", "hello"], ["llm", "hi"]], "query": "how are you?" }
Responses will be returned as dictionaries. Responses should contain the following:
response
- String LLM response to the query
When running this as a docker container, the XDG_CONFIG_HOME
envvar is set to /config
.
A configuration file at /config/neon/diana.yaml
is required and should look like:
MQ:
port: <MQ Port>
server: <MQ Hostname or IP>
users:
<LLM MQ service_name>:
user: <MQ user>
password: <MQ user's password>
LLM_<LLM NAME uppercase>:
num_parallel_processes: <integer > 0>
An LLM may be configured to connect to a /chatbots
vhost and participate in
discussions as described in the chatbots project.
One LLM may define multiple personas to participate as:
llm_bots:
<LLM Name>:
- name: Assistant
description: You are a personal assistant who responds in 40 words or less
- name: Author
description: You are an author and expert in literary history
- name: Student
description: You are a graduate student working in the field of artificial intelligence
enabled: False
LLM Name
is defined in the propertyNeonLLMMQConnector.name