<<<<<<< HEAD
A simple CLI for exporting AI models in multiple formats, quantization schemes, and device targets, with built-in evaluation.
-
Export Formats:
- ONNX
- TorchScript
- Hugging Face Hub
- TensorRT (NVIDIA)
- TFLite (TensorFlow Lite)
- Core ML (Apple)
-
Quantization:
- FP16
- INT8
-
Device Targets:
- CPU
- CUDA
- (extendable to other backends)
-
CLI Scripts:
solo-export— run all exports + evalsolo-export-onnx— ONNX onlysolo-export-ts— TorchScript onlysolo-export-eval— Evaluate exports
-
Evaluation: stub evaluations per export artifact for metrics like accuracy, latency, perplexity
pip install solo-exporterCreate config/export.yaml:
model:
base: gpt2
export:
formats:
- onnx
- torchscript
- hf
- tensorrt
- tflite
- coreml
quantization:
- fp16
- int8
devices:
- cpu
- cuda
output_dir: output
evaluation:
dataset: path/to/eval_dataset.jsonl
metrics:
- accuracy
- latency
- perplexitysolo-export --config config/export.yamlsolo-export-onnx --config config/export.yamlsolo-export-ts --config config/export.yamlsolo-export-eval --config config/export.yamlResults will be saved under output/ in subfolders by format, quantization, and device.
output/
├── gpt2_fp16_onnx_cpu/
│ ├── model.onnx
│ └── metrics.json
├── gpt2_int8_ts_cuda/
│ ├── model.pt
│ └── metrics.json
└── ...
=======
6b862b48f4da0a4e9869226c9334978206af944f