This repo currently contains samples to fine-tune HuggingFace models for text summarization using Microsoft's Azure Machine Learning service. These samples could be adapted to fine-tune models for other NLP tasks or product scenarios.
- AzureML v2 CLI examples for fine-tuning HuggingFace models
- Quickstart ARM Templates for fine-tuning HuggingFace models
- Fine-tuned HuggingFace models & results: https://huggingface.co/linydub
- Benchmarking and carbon accounting with MLflow and Azure Monitor Metrics (performance + resource metrics)
- Interactive data visualization example with Azure Monitor Workbook
- AML v2 CLI inference samples with ONNX Runtime and NVIDIA Triton (AML endpoint & deployment)
- AML v2 CLI end-to-end pipeline samples
- Repository documentation and detailed guide for the samples
- More fine-tuned models and benchmark results
*More details about the project and future plans could be found here.
Directory | Description |
---|---|
cloud |
Cloud-specific configuration code |
docs |
Project docs & images |
examples |
AzureML examples for sample tasks |
These samples showcase various methods to fine-tune HuggingFace models using AzureML. All of the samples include DeepSpeed, FairScale, CodeCarbon, MLflow integrations with no additional setup or code.
All logged training metrics are automatically reported to AzureML and MLflow. CodeCarbon also generates a emissions.csv
file by default inside the outputs folder of the submitted run. To disable a package, ommit it from the environment's conda file.
*Sample script for retrieving and aggregating MLflow and resource usage data will be available next update.
More advanced ARM Templates will be available here.
Fine-tuning samples using AML 2.0 CLI could be found here.
Notebook | Description |
---|
Please file an issue through the repo or email me at [email protected]. Feedback is greatly appreciated 🤗