MCP server for Tangerine (Convo AI assistant backend): https://github.com/RedHatInsights/tangerine-backend
You can run the tangerine-mcp server in a container using Podman or Docker. Make sure you have a valid OpenShift token:
Example configuration for running with Podman:
{
"mcpServers": {
"tangerine": {
"command": "podman",
"args": [
"run",
"-i",
"--rm",
"-e", "TANGERINE_URL",
"-e", "TANGERINE_TOKEN",
"-e", "MCP_TRANSPORT",
"quay.io/maorfr/tangerine-mcp:latest"
],
"env": {
"TANGERINE_URL": "https://tangerine.example.com",
"TANGERINE_TOKEN": "REDACTED",
"MCP_TRANSPORT": "stdio"
}
}
}
}Replace REDACTED with the OpenShift token.