The Enterprise Scale AI Factory
is a plug and play solution that automates the provisioning, deployment, and management of AI projects on Azure with a template way of working.
- Plug and play accelerator for: DataOps, MLOps, LLMOps, enterprise scale environment.
Marry multiple best practices & accelerators:
It reuses multiple existing Microsoft accelerators/landingzone architecture and best practices such as CAF & WAF, and provides an end-2-end experience including Dev,Test, Prod environments.- All
PRIVATE
networking: Private endpoints for all services such as Azure Machine Learning, private AKS cluster, private Container registry, Storage, Azure data factory, Monitoring etc- Both for creating artifacts, training, and inference. To avoid data exfiltration, and have high network isolation
- Docs: Securing Azure Machine Learning & its compute: https://learn.microsoft.com/en-us/azure/machine-learning/how-to-secure-training-vnet?view=azureml-api-1&tabs=instance%2Crequired
- All
Plug-and-play
: Dynamicallly create infra-resources per team, including networking dynamically, and RBAC dynamically- Example of dynamicall: Subnet/IP calculator, ACL permission on the datalake for a project team, services "glued together"
Template way of working & Project way of working:
The AI Factory isproject based
(cost control, privacy, scalability per project) and provides multiple templates besides infrastructure template:DataLake template, DataOps templates, MLOps templates
, with selectable project types.- Sub-purpose:
Same MLOps
- weather data scientists chooses to work from Azure Databricks or Azure Machine Learning` - same MLOps template is used. - Sub-purpose:
Common way of working, common toolbox, a flexible one
: A toolbox with a LAMBDA architecture with tools such as: Azure Datafactory, Azure Databricks, Azure Machine Learning, Eventhubs, AKS
- Sub-purpose:
Enterprise scale & security & battle tested
: Used by customers and partners with MLOps since 2019 (see LINKS) to accelerate the development and delivery of AI solutions, with common tooling & marrying multiple best practices. Private networking (private endpoints), as default.
-
AI factory - setup in 60h (Company: Epiroc)
- End-2-End pipelines for use case: How-to -
AI factory
- Technical BLOG -
Microsoft: AI Factory (CAF/MLOps)
documentation : Machine learning operations - Cloud Adoption Framework | Microsoft Learn
Tehnically, there are two IaC automated project types in the AIFactory: ESML, GenAI. Here they are seen connected to PERSONAS.
Personas is a tool the AIFactory uses to map tools, processes and people, to scale AI organizationally as well.
Personas is used to:
- Find resource gaps, define responsibility, or find redesign needs: If you do not have people in your organization that fit a persona description needed to support a process step, you either need to redesign the architecture, change the process, or onboard new people with that persona. Personas is a good tool to define scope of responsibility
- Education: Mapping personas to specific Azure services in the architecture provides the benefits of offering educational sessions and online courses to upskill within.
- Security & Access: Personas mapped to processes, architectures and services can be used to define which services they need access to in a process.
- Project planning & Interactions Personas mapped to each other can be used see which personas that primarily interacts with each other, to be used to setup sync meetings and project planning.
The 2 project types, lives inside of the AIFactory landingzones.
- There are 3 AIFactory AI landingzones: Dev, Stage, Production, where a project is represented.
- The AIFactory has a default scalabillity to automate the creation of ~200-300 AIFactory projects, in each environment.
- One project is usually assigned to a team of 1-10 people with multiple use cases, but sometimes also to run an isolated use case.
The Documentation is organized around ROLES via Doc series.
Doc series | Role | Focus | Details |
---|---|---|---|
10-19 | CoreTeam |
Governance |
Setup of AI Factory. Governance. Infrastructure, networking. Permissions |
20-29 | CoreTeam |
Usage |
User onboarding & AI Factory usage. DataOps for the CoreTeam's data ingestion team |
30-39 | ProjectTeam |
Usage |
Dashboard, Available Tools & Services, DataOps, MLOps, Access options to the private AIFactory |
40-49 | All |
FAQ |
Various frequently asked questions. Please look here, before contacting an ESML AIFactory mentor. |
It is also organized via the four components of the ESML AIFactory:
Component | Role | Doc series |
---|---|---|
1) Infra:AIFactory | CoreTeam | 10-19 |
2) Datalake template | All | 20-29,30-39 |
3) Templates for: DataOps, MLOps, *LLMOps | All | 20-29, 30-39 |
4) Accelerators: ESML SDK (Python, PySpark), RAG Chatbot, etc | ProjectTeam | 30-39 |
- Based on best & proven practices for organizational scale, across projects.
- Best practice:
CAF/AI Factory
: https://docs.microsoft.com/en-us/azure/cloud-adoption-framework/ready/azure-best-practices/ai-machine-learning-mlops#mlops-at-organizational-scale-ai-factories - Best practice:
Microsoft Intelligent Data Platform
: https://techcommunity.microsoft.com/t5/azure-data-blog/microsoft-and-databricks-deepen-partnership-for-modern-cloud/ba-p/3640280Modern data architecture with Azure Databricks and Azure Machine Learning
: https://docs.microsoft.com/en-us/azure/architecture/solution-ideas/articles/azure-databricks-modern-analytics-architecture
- Best practice:
Datalake design
: https://docs.microsoft.com/en-us/azure/storage/blobs/data-lake-storage-best-practicesDatamesh
: https://martinfowler.com/articles/data-mesh-principles.html- Credit to: Zhamak Dehghani
- Best practice:
- ESML has a default scaling from 1-250 ESMLprojects for its
ESML AI Factory
.- That said, the scaling roof is on IP-plan, and ESML has its own IP-calculator (allocated IP-ranges for 250 is just the default)
Enterprise "cockpit"
over ALL your projects & models.- See what
state
a project are in (Dev,Test,Prod states) withcost dashboard
per project/environment
- See what
Date | Category | What | Link |
---|---|---|---|
2024-03 | Automation | Add core team member | 26-add-esml-coreteam-member.ps1 |
2024-03 | Automation | Add project member | 26-add-esml-project-member.ps1 |
2024-03 | Tutorial | Core-team tutorial | 10-AIFactory-infra-subscription-resourceproviders.md |
2024-03 | Tutorial | End-user tutorial | 01-jumphost-vm-bastion-access.md |
2024-03 | Tutorial | End-user tutorial | 03-use_cases-where_to_start.md |
2024-02 | Tutorial | End-user installation Compute Instance | R01-install-azureml-sdk-v1+v2.m |
2024-02 | Datalake - Onboarding | Auto-ACL on PROJECT folder in lakel | - |
2023-03 | Networking | No Public IP: Virtual private cloud - updated networking rules | https://learn.microsoft.com/en-us/azure/machine-learning/v1/how-to-secure-workspace-vnet?view=azureml-api-1&preserve-view=true&tabs=required%2Cpe%2Ccli |
2023-02 | ESML Pipeline templates | Azure Databricks: Training and Batch pipeline templates. 100% same support as AML pipeline templates (inner/outer loop MLOps) | - |
2022-08 | ESML infra (IaC) | Bicep now support yaml as well | - |
2022-10 | ESML MLOps | ESML MLOps v3 advanced mode, support for Spark steps ( Databricks notebooks / DatabrickStep ) | - |
ESML stands for: Enterprise Scale ML.
This accelerator was born 2019 due to a need to accelerated DataOps and MLOps.
The accelerateor was then called ESML, We now only call this acceleration ESML, or project type=ESML, in the Entperise Scale AIFActory
Innovating with AI and Machine Learning, multiple voices expressed the need to have an Enterprise Scale AI & Machine Learning Platform
with end-2-end
turnkey DataOps
and MLOps
.
Other requirements were to have an enterprise datalake design
, able to share refined data across the organization
, and high security
and robustness: General available technology only, vNet support for pipelines & data with private endpoints. A secure platform, with a factory approach to build models.
Even if best practices exists, it can be time consuming and complex
to setup such a AI Factory solution
, and when designing an analytical solution a private solution without public internet is often desired since working with productional data from day one is common, e.g. already in the R&D phase. Cyber security around this is important.
Challenge 1:
Marry multiple, 4, best practicesChallenge 2:
Dev, Test, Prod Azure environments/Azure subscriptionsChallenge 3:
Turnkey: Datalake, DataOps, INNER & OUTER LOOP MLOps Also, the full solution should be able to be provisioned 100% viainfrastructure-as-code
, to be recreated and scale across multiple Azure subscriptions, andproject-based
to scale up to 250 projects - all with their own set of services such as their own Azure machine learning workspace & compute clusters.
To meet the requirements & challenge, multiple best practices needed to be married and implemented, such as: CAF/WAF, MLOps, Datalake design, AI Factory, Microsoft Intelligent Data Platform / Modern Data Architecture.
An open source initiative could help all at once, this open-source accelerator Enterprise Scale ML(ESML) - to get an AI Factory on Azure
ESML
provides an AI Factory
quicker (within 4-40 hours), with 1-250 ESMLProjects, an ESML Project is a set of Azure services glued together securely.
Challenge 1 solved:
Marry multiple, 4, best practicesChallenge 2 solved:
Dev, Test, Prod Azure environments/Azure subscriptionsChallenge 3 solved:
Turnkey: Datalake, DataOps, INNER & OUTER LOOP MLOpsESML marries multiple best practices
into onesolution accelerator
, with 100% infrastructure-as-code
The below is how it looked like, when ESML automated both the infrastructire, and generating Azure machine learning pipelines, with 3 lines of code.
TRAINING & INFERENCE pipeline templates types in ESML AIFactory that accelerates for the end-user.
- 0.1% percentage of the code to write, to go from R&D process, to productional Pipelines:
This repository is a push-only mirror. Ping Joakim Åström for contributions / ideas.
Since "mirror-only" design, Pull requests are not possible, except for ESML admins. See LICENCE file (open source, MIT license)
Speaking of open source, contributors:
- Credit to
Kim Berg
andBen Kooijman
for contributing! (kudos to the ESML IP calculator and Bicep additions for esml-project type) - Credit to
Christofer Högvall
for contributing! (kudos to the Powershell script, to enable Resource providers, if not exits)azure-enterprise-scale-ml\environment_setup\aifactory\bicep\esml-util\26-enable-resource-providers.ps1