All Tags
AWS
ai
algorithm-design
architecture
browser
cloud
cloud-efficiency
cloud-principles
cost-reduction
data-centric
data-compression
data-processing
deployment
design
documentation
edge-computing
email-sharing
energy-efficiency
energy-footprint
enterprise-optimization
green-ai
hardware
libraries
llm
locality
machine-learning
maintainability
management
measured
microservices
migration
mobile
model-optimization
model-training
multi-objective
network-traffic
parameter-tuning
performance
queries
rebuilding
scaling
services
storage-optimization
strategies
tabs
template
testing
workloads
Tactic: Use Subset-Based Training
Tactic sort:
Awesome Tactic
Type: Architectural Tactic
Category: green-ml-enabled-systems
Title
Use Subset-Based Training
Description
Subset-Based Training involves training models on disjoint subsets of data rather than whole dataset training. Energy used for training is significantly lowered and final ensemble accuracy is relatively unchanged.
Participant
Machine Learning Practitioners and Researchers.
Related software artifact
Machine Learning Training Data.
Context
Machine Learning Ensemble. Green AI.
Software feature
Horizontal Partitioning.
Tactic intent
Reduce energy used in model training by subdividing the dataset and using horizontal partitioning.
Target quality attribute
Energy Efficiency.
Other related quality attributes
Accuracy.
Measured impact
Whole-dataset training consumed on average 45.7% more energy, but offered a negligible 0.0095 increase in accuracy compared to subset-based training. Energy measured in Joule (J) for training all base models, the optional meta-model, and fusion-time inference. Accuracy measured in F1.
