Skip to content

Apache 2.0-licensed open source operations stack for private AI inference with open models. Run LLMs (7B-70B) locally with vLLM, OpenAI-compatible API, web dashboard, chat UI, admin panel, and hardware monitoring.

License

Notifications You must be signed in to change notification settings

scalytics/ScalyticsConnect

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

24 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Scalytics Connect

CI Secrets Scan

Scalytics Connect is a private AI operations platform for organizations that need governance, deployment control, and auditability for GenAI workloads.

Built by Scalytics, it is designed for teams that cannot rely on unmanaged black-box AI tooling.

The Niche We Serve

Most AI chat stacks optimize for speed-to-demo. Scalytics Connect optimizes for production control:

  • data locality and controlled model access
  • role-based permissions and policy enforcement
  • local + external model support in one platform
  • self-hosted deployment workflows (including GCP)

For Enterprise Teams

Use Scalytics Connect when you need:

  • governed AI operations across departments
  • user/group/permission controls at model and provider level
  • API key and integration governance
  • operational visibility (usage, maintenance, admin controls)
  • deployment ownership for compliance and security requirements

For Developers

What you get in this repo:

  • Node/Express backend with modular routes/services
  • React frontend for chat + administration
  • Python services/workers for inference and deep research
  • vLLM-based local inference support
  • MCP tool and agent integration surface

Key areas:

  • src/routes/ API surface
  • src/services/ orchestration, providers, model logic
  • src/python_services/ deep-search/research services
  • frontend/ web app
  • deploy/ deployment scripts
  • docs/ OpenAPI/docs assets

Core Capabilities

  • Multi-model routing (local and external providers)
  • Local model lifecycle management (discover, download, activate, optimize)
  • User/group/permission management
  • Provider/API key administration
  • Agent workflows + MCP tools
  • Deep-search/research pipeline

Quick Start (Local)

Backend:

npm install
npm run setup
npm run dev

Frontend (separate process):

cd frontend
npm install
npm start

Production/GCP deployment guidance: deploy/README.md

Documentation

Start here: docs/index.md

  • User docs: docs/user/
  • Admin docs: docs/admin/
  • Developer docs: docs/developer/
  • OpenAPI spec: docs/openapi.json
  • Static API docs: docs/api-docs.html

Contributing and Governance

  • Contribution guide: CONTRIBUTING.md
  • Code of conduct: CODE_OF_CONDUCT.md
  • Security reporting: SECURITY.md
  • Maintainer policy: MAINTAINERS.md
  • Changelog: CHANGELOG.md

API and Docs

  • OpenAPI spec: docs/openapi.json
  • Build static API docs:
npm run build:api-docs

Security Baseline

Before production use:

  • configure secrets only via environment variables
  • run services with least privilege
  • validate access policies for users, groups, models, and providers
  • review deployment hardening in deploy/

License

Apache License 2.0. See LICENSE and NOTICE.

About

Apache 2.0-licensed open source operations stack for private AI inference with open models. Run LLMs (7B-70B) locally with vLLM, OpenAI-compatible API, web dashboard, chat UI, admin panel, and hardware monitoring.

Topics

Resources

License

Code of conduct

Contributing

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •