Vision Transformer (ViT) models for image classification converted to ggml format
| Model | Disk | Mem | SHA | 
|---|---|---|---|
| tiny | 12 MB | ~20 MB | 25ce65ff60e08a1a5b486685b533d79718e74c0f | 
| small | 45 MB | ~52 MB | 7a9f85340bd1a3dcd4275f46d5ee1db66649700e | 
| base | 174 MB | ~179 MB | a10d29628977fe27691edf55b7238f899b8c02eb | 
| large | 610 MB | ~597 MB | 5f27087930f21987050188f9dc9eea75ac607214 | 
The models are pre-trained on ImageNet21k then finetuned on ImageNet1k with a patch size of 16 and an image size of 224.
For more information, visit:
- Downloads last month
 - 72
 
							Hardware compatibility
						Log In
								
								to view the estimation
16-bit
	Inference Providers
	NEW
	
	
	This model isn't deployed by any Inference Provider.
	๐
			
		Ask for provider support