# llm-smollm2 **Repository Path**: tom_ud/llm-smollm2 ## Basic Information - **Project Name**: llm-smollm2 - **Description**: No description available - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2025-02-25 - **Last Updated**: 2025-02-25 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # llm-smollm2 [![PyPI](https://img.shields.io/pypi/v/llm-smollm2.svg)](https://pypi.org/project/llm-smollm2/) [![Changelog](https://img.shields.io/github/v/release/simonw/llm-smollm2?include_prereleases&label=changelog)](https://github.com/simonw/llm-smollm2/releases) [![Tests](https://github.com/simonw/llm-smollm2/actions/workflows/test.yml/badge.svg)](https://github.com/simonw/llm-smollm2/actions/workflows/test.yml) [![License](https://img.shields.io/badge/license-Apache%202.0-blue.svg)](https://github.com/simonw/llm-smollm2/blob/main/LICENSE) SmolLM2-135M-Instruct.Q4_1 for LLM. [Background on this project](https://simonwillison.net/2025/Feb/7/pip-install-llm-smollm2/). ## Installation Install this plugin in the same environment as [LLM](https://llm.datasette.io/). ```bash llm install llm-smollm2 ``` If you have [uv](https://github.com/astral-sh/uv) installed you can chat with the model without any installation step like this: ```bash uvx --with llm-smollm2 llm chat -m SmolLM2 ``` ## Usage This plugin bundles a full copy of the [SmolLM2-135M-Instruct.Q4_1](https://huggingface.co/QuantFactory/SmolLM2-135M-Instruct-GGUF/blob/ab810cf68114990406fdf996510dd3d3c6adbdf5/SmolLM2-135M-Instruct.Q4_1.gguf) quantized version of the [SmolLM2-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct) model by [HuggingFaceTB](https://huggingface.co/HuggingFaceTB). Once installed, run the model like this: ```bash llm -m SmolLM2 'Are dogs real?' ``` Or to chat with the model (keeping it resident in memory): ```bash llm chat -m SmolLM2 ``` ## Development To set up this plugin locally, first checkout the code. Then create a new virtual environment: ```bash cd llm-smollm2 python -m venv venv source venv/bin/activate ``` Now install the dependencies and test dependencies: ```bash llm install -e '.[test]' ``` To run the tests: ```bash python -m pytest ```