--- tags: - neuron - optimized - aws-neuron - text-generation base_model: HuggingFaceTB/SmolLM2-135M-Instruct --- # Neuron-Optimized HuggingFaceTB/SmolLM2-135M-Instruct This repository contains AWS Neuron-optimized files for [HuggingFaceTB/SmolLM2-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct). ## Model Details - **Base Model**: [HuggingFaceTB/SmolLM2-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct) - **Task**: text-generation - **Optimization**: AWS Neuron compilation - **Generated by**: [badaoui](https://huggingface.co/badaoui) - **Generated using**: [Optimum Neuron Compiler Space](https://huggingface.co/spaces/optimum/neuron-export) ## Usage This model has been optimized for AWS Neuron devices (Inferentia/Trainium). To use it: ```python from optimum.neuron import NeuronModelForCausalLM model = NeuronModelForCausalLM.from_pretrained("badaoui/HuggingFaceTB-SmolLM2-135M-Instruct-neuron") ``` ## Performance These files are pre-compiled for AWS Neuron devices and should provide improved inference performance compared to the original model when deployed on Inferentia or Trainium instances. ## Original Model For the original model, training details, and more information, please visit: [HuggingFaceTB/SmolLM2-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct)