Skip to content

VIAcode/AICore

Repository files navigation

AICore Overview

License: MIT

AI Core

The AI Core is an open-source toolkit that streamlines the development, deployment, and management of AI-based applications. Whether you’re leveraging Azure AI, Azure OpenAI, or custom models, this platform provides an all-in-one solution. With built-in support for agent workflows, session management, user management, and cost tracking, the platform ensures a secure and scalable environment for all your AI needs. The platform is available as a source code and a Docker image for seamless integration.

💡 Key Features

  • AI Models Integration: Manage different types of AI models effortlessly, with built-in support for Azure Open AI, Azure AI, Azure AI Document Intelligence and custom models.
  • Agents & Flows: Enable complex workflows and AI-driven tasks with customizable agents and composite workflows.
  • RAG skills: link with data sources and use retrieval-augmented generation to gather information
  • User Management: Manage users and groups securely, with support for Single Sign-On (SSO) and Microsoft Entra ID integration.
  • Cost Control: Gain full visibility and control over AI-related costs by managing usage across users, models, and agents.
  • AI Jobs Scheduler: Schedule and run background agents or workflows, automating repetitive tasks to improve efficiency.

🤖 Agents

AI Core includes agents, modular components that enhance your AI workflows by automating specific tasks or integrating third-party services. Agents act as independent processes within the system, facilitating functions such as:

  • Data preprocessing or enrichment
  • Triggering external APIs
  • Scheduling tasks or recurring jobs
  • Supporting domain-specific operations

Developers can create custom agents or use predefined ones available in the platform to tailor solutions to specific needs. This modular design ensures flexibility and reduces development time for complex AI-driven applications.

Full Agents Guide

🪄 Available LLMs

Large Language Models (LLMs) are advanced AI models trained on vast amounts of text data to understand and generate human-like language. They can perform a variety of tasks, including answering questions, summarizing documents, writing code, and engaging in conversational interactions. LLMs are essential for building intelligent, language-based applications such as chatbots, virtual assistants, and content generation tools.

AI Core currently supports the following LLMs:

  • GPT-4o-mini: A lightweight version optimized for fast inference and low-resource environments, making it ideal for quick-response scenarios with minimal infrastructure requirements.
  • GPT-4o: A more powerful model offering greater accuracy and broader language capabilities, suitable for demanding tasks like advanced natural language understanding and summarization.

With these models, AI Core ensures a balanced approach, providing options for both performance-efficient and high-accuracy solutions, meeting a variety of project needs.

💬 Chat

Playground AI Core offers a Chat Area for data handling and models testing where users can:

  • Experiment with data inputs and outputs, gaining insights into model behavior in real time.
  • Test various models and configurations.
  • Simulate workflows, ensuring that agents, data flows, and AI models work together as intended.
  • Visualize key metrics such as response times, accuracy, and cost impact of various AI tasks.
  • Chat provides a safe, user-friendly environment to prototype ideas quickly and debug workflows before full-scale deployment.

AI Core API References

Deployment

This guide explains how to deploy AI Core using the pre-built Docker images. Follow these steps to get the solution up and running efficiently.


Prerequisites

Before deploying, ensure you have the following installed and configured on your system:

  1. Docker:
  2. Docker Hub Account (Optional):
    • If the Docker images are private, log in to Docker Hub using your credentials:
      docker login

Steps to Deploy the Solution

1. Pull the Docker Images

The pre-built Docker images for AI Core are hosted on Docker Hub. To download the images, run:

Ingestion Service Image

docker pull viacode/ai-core-file-ingestion:latest

API Service Image

docker pull viacode/ai-core:latest

2. Create Environment Configuration (Optional)

Some configurations may require environment variables. Create a .env file in your project directory to define these variables:

APP_ENV=production
APP_PORT=8080
DB_HOST=your-database-host
DB_USER=your-database-user
DB_PASSWORD=your-database-password
INGESTION_PORT=5000
API_PORT=8081

Ensure you replace the placeholders with actual values.


3. Run the Docker Containers

3.1 Start the Ingestion Service Container

Start the ingestion service container using the following command:

docker run -d \
  --name ai-core-file-ingestion \
  -p 8080:8080 \
  --env-file .env \
  viacode/ai-core-file-ingestion:latest

3.2 Start the API Service Container

Start the API service container using the following command:

docker run -d \
  --name ai-core \
  -p 8081:8081 \
  --env-file .env \
  viacode/ai-core:latest

4. Verify the Deployment

To ensure the containers are running, execute:

docker ps

You should see both containers listed. Verify the logs to ensure everything is working correctly:

Ingestion Service Logs

docker logs ai-core-file-ingestion

API Service Logs

docker logs ai-core

Access the services in your browser:

  • Ingestion Service: http://localhost:8080
  • API Service: http://localhost:8081

5. Manage the Docker Containers

Here are some useful commands to manage the containers:

  • Stop the containers:

    docker stop ai-core-file-ingestion ai-core
  • Restart the containers:

    docker restart ai-core-file-ingestion ai-core
  • Remove the containers:

    docker rm ai-core-file-ingestion ai-core
  • Remove the Docker images (if needed):

    docker rmi viacode/ai-core-file-ingestion:latest viacode/ai-core:latest

Happy deploying!

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •