πΏ Sisigoks/FloraSense
FloraSense is a fine-tuned Vision Transformer (ViT) model designed for accurate classification of plant species and flora-related imagery. It builds on top of the powerful google/vit-base-patch16-224 base model and is fine-tuned on the Planter_GARDEN_EDITION dataset curated by Sisigoks, which includes over 10,000 diverse plant images.
π§ Model Description
- Architecture: Vision Transformer (ViT)
- Base Model:
google/vit-base-patch16-224 - Task: Image Classification
- Use Case: Automated plant and flora species recognition in digital botany, garden classification systems, plant care apps, biodiversity projects, and educational tools.
π Model Performance
- Evaluation Accuracy: 35.46%
- Evaluation Loss: 4.2894
- Epochs Trained: 10
- Evaluation Speed:
- 33.9 samples/sec
- 2.12 steps/sec
β οΈ While the accuracy may appear moderate, the model is handling over 10,000 highly similar plant species, making this a non-trivial challenge in fine-grained classification.
π§ͺ Training Procedure
| Hyperparameter | Value |
|---|---|
| Learning Rate | 5e-5 |
| Train Batch Size | 16 |
| Eval Batch Size | 16 |
| Gradient Accumulation | 4 |
| Total Effective Batch | 64 |
| Optimizer | Adam (Ξ²1=0.9, Ξ²2=0.999) |
| Scheduler | Linear w/ warmup (10%) |
| Epochs | 15 |
| Seed | 42 |
- Framework: PyTorch
- Libraries: Transformers 4.45.1, Datasets 3.0.1, Tokenizers 0.20.0
π Dataset
- Name:
Sisigoks/Planter_GARDEN_EDITION - Type: Image Classification
- Language: English
- Scope: Over 10,000 unique plant and floral species
- Format: Real-world garden and nature photography
- Use Case: Realistic and diverse training scenarios for classification models
β Intended Use
Use Cases
- Botanical image recognition apps
- Educational tools for students and researchers
- Smart gardening & plant care solutions
- Field-use flora identification via AR and mobile apps
Target Users
- Botanists
- AI and ML researchers
- Gardeners and farmers
- Biology educators and students
β οΈ Limitations
- May confuse visually similar species due to fine-grained class diversity.
- Performance could degrade in poor lighting or occlusion-heavy environments.
- Biases may exist based on the geographic scope of the dataset (e.g., underrepresentation of tropical or rare plants).
π Ethical Considerations
- Accuracy: Misclassification of medicinal/toxic plants can have real-world safety implications.
- Bias: Regional, lighting, or season-specific training data may skew predictions in certain environments.
- Usage: This is a research-grade model and should not be relied on for critical decisions without expert validation.
π How to Use
from transformers import AutoImageProcessor, AutoModelForImageClassification
from PIL import Image
import torch
# Load model and processor
processor = AutoImageProcessor.from_pretrained("Sisigoks/FloraSense")
model = AutoModelForImageClassification.from_pretrained("Sisigoks/FloraSense")
# Load and preprocess image
image = Image.open("your_image.jpg")
inputs = processor(images=image, return_tensors="pt")
# Inference
with torch.no_grad():
outputs = model(**inputs)
logits = outputs.logits
predicted_label = logits.argmax(-1).item()
print(f"Predicted class ID: {predicted_label}")
π Citation
If you use this model or dataset in your work, please cite:
@misc{sisigoks_florasense_2025,
author = {Sisigoks},
title = {FloraSense: ViT-based Fine-Grained Plant Classifier},
year = {2025},
publisher = {Hugging Face},
howpublished = {\url{https://huggingface.co/Sisigoks/FloraSense}}
}
π Acknowledgements
- Hugging Face π€ β for providing the model and dataset hosting infrastructure.
- Google Research β for the original ViT architecture that enabled scalable vision transformers.
- Downloads last month
- 369
Model tree for Sisigoks/FloraSense
Base model
google/vit-base-patch16-224