Introduction
This repository is part of playbook for experiments on fine tuning small language models using LoRA, exporting them to ONNX and running them locally using ONNX compatibale runtime like javascript(node js) and WASM (browser)
Before you start
- Clone the repository https://github.com/code2k13/onnx_javascript_browser_inference
- Copy all files from this repository to the
model_filesdirectory of the cloned github repository. - Run
npm install
To run NodeJS example (NodeJS + onnx-runtime, server side)
- Simple run
node app.jsThis is what you should see
To run web browser based demo (WASM based in-browser inference)
- Simply access
web.htmlfrom a local server (examplehttp://localhost:3000/web.html)
This is what you should see
Citation
@misc{allal2024SmolLM,
title={SmolLM - blazingly fast and remarkably powerful},
author={Loubna Ben Allal and Anton Lozhkov and Elie Bakouch and Leandro von Werra and Thomas Wolf},
year={2024},
}
- Downloads last month
- 58
Model tree for code2022/SmolLM2-135M-Instruct-Paraphrase
Base model
HuggingFaceTB/SmolLM2-135M
Quantized
HuggingFaceTB/SmolLM2-135M-Instruct



