HRM-Text1
HRM-Text1 is an experimental instruction-following text generation model based on the Hierarchical Recurrent Memory (HRM) architecture. It is trained on the databricks/databricks-dolly-15k dataset, which consists of instruction–response pairs across multiple task types.
The model utilizes the HRM structure, consisting of a "Specialist" module for low-level processing and a "Manager" module for high-level abstraction and planning. This architecture aims to handle long-range dependencies more effectively by summarizing information at different temporal scales.
Model Description
- Architecture: Hierarchical Recurrent Memory (HRM)
 - Training Data: databricks/databricks-dolly-15k
 - Original Paper: Hierarchical Reasoning Model
 - Tokenizer: 
t5-small(slow T5 SentencePiece) - Vocab Size: 32100
 - Objective: Causal Language Modeling
 
Latest Performance (Epoch 20)
- Validation Loss: 
3.6668 - Validation Perplexity: 
39.13 
	Inference Providers
	NEW
	
	
	This model isn't deployed by any Inference Provider.
	🙋
			
		Ask for provider support
Model tree for Viharikvs/HRM-Text1-UltraChat
Base model
google-t5/t5-small