-
Notifications
You must be signed in to change notification settings - Fork 37
2.3.16 Satellite: omnichain
Handle:
omnichain
URL: http://localhost:34081
Efficient visual programming for AI language models
# [Optional] pre-build the image
harbor build omnichain
# Start the service
harbor up omnichain
# [Optional] Open the UI
harbor open omnichain
By default, Omnichain is not pre-configured with any external connections, as everything is specific for any given chain. It means that you'd connect specific chains to the LLM backends in Harbor upon creating them.
Harbor runs a custom version of the omnichain
that was made compatible with webui
. You'll see your chains as models in your Workplace there.
When running, Harbor will mount a workspace folder to the container. You can access it with:
# Open omnichain workspace in the file manager
harbor omnichain workspace
# See where the workspace is located,
# the folder is relative to $(harbor home)
harbor config get omnichain.workspace
In the workspace, following folders are used:
-
custom_nodes
- will provide files for the Custom Nodes functionality -
data
-omnichain
persistence: chains, avatars, etc. -
files
- fill be available as/app/omnichain/files
in your chains, that can be used with File nodes
Please refer to official documentation for more information:
To make the testing easier, you can use a sample chain that was pre-configured to work with ollama
in Harbor.
- HarborChat.json - chat with Harbor CLI