Frameworks: PyTorch (default), Hugging Face Transformers, scikit-learn, XGBoost, LightGBM, CatBoost, Keras / TensorFlow when required
Fine-tuning: Hugging Face TRL, Axolotl, Unsloth, PEFT (LoRA, QLoRA), DeepSpeed, FSDP, vLLM for serving
Computer vision: YOLOv8/v10, Ultralytics, MMDetection, Detectron2, SAM 2, OpenMMLab, OpenCV, Roboflow for dataset ops
Experiment tracking: MLflow, Weights & Biases, Neptune, Comet, ClearML — versioned datasets, models, and runs
Data & feature stores: Pandas, Polars, DuckDB, Dask, Spark; Feast, Tecton for feature stores; DVC for dataset versioning
Training infrastructure: AWS SageMaker, Vertex AI, Azure ML, Modal, Lambda Labs, RunPod, Paperspace, on-prem A100/H100
Inference serving: BentoML, Triton Inference Server, TorchServe, vLLM, TGI, Modal, AWS Lambda + container, SageMaker endpoints
Edge / on-device: Core ML, TFLite, ONNX Runtime, MLC LLM, Llama.cpp for mobile and embedded inference
MLOps & monitoring: Evidently AI, Arize, WhyLabs, Fiddler — for drift, data quality, and bias monitoring in production
Eval & responsible AI: Custom eval harnesses, model cards (Hugging Face / Google), fairness metrics broken down by demographic segments