A profiling framework for analyzing ExecuTorch model performance with SME2 acceleration. Provides operator-level performance analysis to identify bottlenecks and measure SME2 acceleration impact.
This repository provides a profiling framework with:
- Model Export Tools: Export PyTorch models to ExecuTorch
.pteformat with XNNPACK backend delegation - Profiling Pipeline: Automated performance measurement pipeline for macOS and Android platforms
- SME2-on and SME2-off comparison runs
- Timing-only runs for accurate latency measurements
- Trace-enabled runs for kernel-level analysis
- Automatic ETDump analysis and CSV generation
- Analysis Tools: Operator-level performance analysis and bottleneck identification
- Operator-category breakdown (CONV, GEMM, Data Movement, Elementwise, Other)
- Operator-specific bottleneck analysis
- Portable vs delegated operator identification
- Kernel-level insights (SME2 vs standard kernels)
- Report Generation: Markdown reports with performance analysis
- Agent Skills: Automation skills for AI coding assistants
- 8 skills covering setup, build, export, profiling, analysis, and reporting
- Workflows that can be automated or run manually
- Model Onboarding Scaffolding: Framework for adding new models to the profiling workflow
- Example Models: Reference implementations (toy_cnn, mobilenet_v3_small, EdgeTAM)
This framework provides:
- Export Models: Convert PyTorch models to ExecuTorch
.pteformat with proper backend delegation - Measure Performance: Run models with SME2 acceleration on and off to measure speedup
- Analyze Bottlenecks: Break down inference time by operator categories to identify where time is spent
- Identify Optimization Opportunities: Discover which operators benefit from SME2 and which become new bottlenecks (often data movement operations)
- Generate Reports: Create reports with performance analysis for optimization
Key Insight: After SME2 accelerates CONV and GEMM operations (3-15× faster), data movement operations (transpose, reshape, layout conversions) often become the dominant bottleneck. This framework makes the bottleneck shift visible, showing where to focus optimization efforts.
This code repository accompanies the Profiling ExecuTorch Models with SME2 on Arm learning path, which provides additional documentation and step-by-step instructions.
-
Clone this repository:
git clone https://github.com/ArmDeveloperEcosystem/sme-executorch-profiling.git executorch_sme2_kit cd executorch_sme2_kit -
Set up ExecuTorch:
bash model_profiling/scripts/setup_repo.sh
-
Build runners:
bash model_profiling/scripts/build_runners.sh
-
Activate venv and export a model:
source .venv/bin/activate python model_profiling/export/export_model.py \ --model <model_name> \ --dtype fp16 \ --outdir out_<model>/artifacts/
-
Create config and run profiling pipeline:
# Copy template cp model_profiling/configs/templates/mac_template.json \ model_profiling/configs/my_experiment.json # Edit config: set "model" to your .pte path # Edit config: set "output_root" to "out_<model>/runs/mac" python model_profiling/scripts/mac_pipeline.py \ --config model_profiling/configs/my_experiment.json # Pipeline automatically runs analysis and generates CSV files
-
View results:
# Analysis runs automatically during pipeline execution # Results include CSV files, pipeline_summary.json/md, and analysis_summary.json # Optional: Re-run analysis if needed python model_profiling/scripts/analyze_results.py \ --run-dir out_<model>/runs/mac # Generate markdown report (base report) python model_profiling/scripts/generate_report.py \ --run-dir out_<model>/runs/mac # Operator-specific bottleneck analysis python model_profiling/tools/analyze_etdump_csv.py \ --timeline-csv out_<model>/runs/mac/<experiment>/*_all_runs_timeline.csv \ --compare out_<model>/runs/mac/<experiment_off>/*_all_runs_timeline.csv \ --name1 "SME2-Off" \ --name2 "SME2-On" \ --output-dir out_<model>/runs/mac/ \ --verbose
Note: The base report shows category-level breakdown. For operator-level bottlenecks and portable vs delegated analysis, use
analyze_etdump_csv.py. See agent skill07_report_generation.mdfor the workflow.
This repository includes agent skills in agent_skill_ml_profiling/ for AI coding assistants (Claude, Cursor, Copilot, etc.) and CI pipelines to automate the profiling workflow.
What are agent skills? These are self-contained automation playbooks with:
- Clear prerequisites and verification steps
- Ordered, executable commands
- Expected outputs and success criteria
- Failure handling and troubleshooting guidance
Available Skills:
01_setup_workspace.md- Initialize profiling environment (~30 min)02_build_runners.md- Build SME2-on/off runner binaries (~20 min)03_export_model.md- Export PyTorch model to ExecuTorch .pte (~5 min)04_run_profiling.md- Execute profiling pipeline (~10 min)05_analyze_results.md- Generate operator-category breakdown (~2 min)06_validate_workflow.md- End-to-end smoke test (~15 min)07_report_generation.md- Generate markdown report (~1 min)08_onboard_edgetam.md- Onboard EdgeTAM image encoder model (~5 min)
Quick Start with Agent Skills:
- For AI assistants: Reference skills by name when automating profiling tasks
- For developers: Use skills as step-by-step playbooks (run commands sequentially, verify each step)
- For CI/CD: Chain skills together for automated regression testing
See agent_skill_ml_profiling/readme.md for the skill catalog and usage patterns.
model_profiling/export/- Model export script (with registry patching)model_profiling/models/- Model registry and onboarding scaffoldingmodel_profiling/scripts/- Pipeline scripts (mac, android, analysis, setup, build)model_profiling/configs/- Configuration templates and examplesmodel_profiling/tools/- Analysis tools (ETDump to CSV, bottleneck analysis)model_profiling/pipeline/- Core pipeline orchestration codeagent_skill_ml_profiling/- Agent skills for automation (8 skills)out_<model>/artifacts/- Exported.ptefiles (created during export)out_<model>/runs/- Profiling results (created during pipeline runs)
Note: Replace <model> with your actual model name. The out_<model>/ directories are created automatically when you export and run profiling.
The profiling pipeline is model-agnostic - once you export a .pte file, the same commands work for any model. Only the model export step is model-specific.
You have two options for detailed onboarding instructions:
-
Agent Skill (recommended for step-by-step automation): See
agent_skill_ml_profiling/08_onboard_edgetam.mdfor an EdgeTAM onboarding example that demonstrates:- Wrapper classes for input/output normalization
- Operator replacement strategies
- Shape constraint handling
- Export-friendly refactoring patterns
Note: EdgeTAM is a third-party open source project. When cloning EdgeTAM, you must maintain all copyright notices and comply with EdgeTAM's license terms. See the onboarding skill for details.
-
Learning Path: See the learning path documentation for onboarding instructions and tutorials.
-
Create
model_profiling/models/<your_model>/with:__init__.py- Registers the modelmodel.py- ImplementsEagerModelBaseinterfacevendor/- (optional) vendored upstream code
-
Export using the same
export_model.pyscript:python model_profiling/export/export_model.py \ --model <your_model> \ --dtype fp16 \ --outdir out_<your_model>/artifacts/
-
Run the same pipeline scripts with your exported
.pte:python model_profiling/scripts/mac_pipeline.py \ --config model_profiling/configs/my_experiment.json
The pipeline automatically handles analysis and report generation - no model-specific changes needed after export.
- Agent Skills:
agent_skill_ml_profiling/readme.md- Catalog of automation skills - Command Reference:
model_profiling/pipeline_commands.md- Detailed workflow commands - Scripts Overview:
model_profiling/scripts/readme.md- Script documentation - Report Generation:
agent_skill_ml_profiling/07_report_generation.md- Workflow for generating reports with operator-specific bottleneck analysis, portable vs delegated operator identification, and kernel-level insights - Model Onboarding:
agent_skill_ml_profiling/08_onboard_edgetam.md- Step-by-step EdgeTAM onboarding example
- Learning Path: Profiling ExecuTorch Models with SME2 on Arm - Additional documentation and tutorials (optional - this repository is self-contained)
This project is licensed under the Apache License, Version 2.0. See the LICENSE file for details.
Third-party licenses: This repository includes instructions to use third-party open source projects (e.g., EdgeTAM). When using such projects, you must comply with their respective licenses. See the EdgeTAM onboarding documentation for license compliance requirements.