blog photo

My Stack: Tools, Frameworks, and Python Packages I Use for AI and ML 💻

Building an AI startup in 2025 means making sharp choices with your tech stack. Every tool and service needs to be lightweight, scalable, and developer-friendly—especially when you’re running lean.

In this post, I'm breaking down my core ML stack, powered by Azure Cloud Services and Mask R-CNN as my primary model architecture.

☁️ Why Azure?

Azure offers a tightly integrated ecosystem for ML workflows—from data storage and orchestration to model training, deployment, and inference. I chose Azure for its maturity, support for Kubernetes, and integration with enterprise services like Azure Active Directory, Key Vault, and Azure DevOps.

🧠 Overview of My Stack

  • Language: Python 3.11+
  • Core ML Framework: PyTorch
  • Main Model: Mask R-CNN (via Detectron2)
  • Labeling Tools: Roboflow, CVAT
  • Training Pipeline: Azure Databricks + Azure ML
  • Storage: Azure Blob Storage, Azure Data Lake Gen2
  • Model Serving: Azure Kubernetes Service (AKS) + Azure ML Model Registry
  • API Gateway: Azure API Management
  • Trigger + Orchestration: Azure Functions, Logic Apps, Event Grid
  • LLM Integration: Azure OpenAI + LangChain
  • Reporting: Azure App Service, PDF via Azure Functions, Power BI Dashboards

🧱 Core Tools & Libraries

  • PyTorch – base framework
  • Detectron2 – for Mask R-CNN and vision models
  • scikit-learn – evaluation metrics
  • albumentations – for image augmentations
  • torchvision, openCV, matplotlib, pycocotools

📦 Azure-Specific Components

Azure Blob Storage + Data Lake

Used to store and organize both raw and preprocessed image data. Perfect for managing large volumes of labeled datasets and experiment outputs.

Azure Databricks

I use this for Spark-based ETL and preprocessing, especially for preparing training datasets at scale.

Azure Machine Learning & Model Registry

My training jobs and experiments run here. Once models are validated, they're registered and deployed using AKS for scalable inference.

Azure Kubernetes Service + API Management

Trained models are containerized and served through AKS, exposed via a REST API, and managed securely with Azure API Management.

Azure Functions + Logic Apps + Event Grid

These handle triggers, automation, and orchestration—such as post-inference workflows or LLM Q&A invocations.

Azure OpenAI + LangChain

I integrate LLMs for post-classification reasoning, auto-report generation, and insights by connecting LangChain with Azure OpenAI endpoints.

Reporting & Output

All generated content flows into Azure App Services (for UI), Power BI dashboards, or is exported to PDFs via serverless Azure Functions.

💬 Final Thoughts

This stack reflects what I believe every AI founder should aim for: speed, modularity, and production readiness.

No hype tools, no overkill infrastructure—just the right services to build, learn, and iterate quickly in production-like environments.

Coming soon: A deep dive into how I annotate data for Mask R-CNN, and my automated training loop using Azure ML pipelines.

Questions or want a code snippet? Let’s connect.

×

About my mission

Partner with me on a project

Copyright © 2025 PixIntel - Designed by Abdul Koomson.