All Tags
AWS
algorithm-design
architecture
cloud-principles
cost-reduction
data-centric
data-compression
data-processing
deployment
design
edge-computing
energy-footprint
hardware
libraries
locality
machine-learning
management
measured
migration
model-optimization
model-training
performance
queries
rebuilding
scaling
services
strategies
template
workloads
Tactic: Apply Sampling Techniques
Tactic sort:
Awesome Tactic
Type: Architectural Tactic
Category: green-ml-enabled-systems
Title
Apply Sampling Techniques
Description
The size of input data has a positive correlation with the energy consumption of computing. Therefore, reducing the size of input data can have a positive impact on energy-efficiency of machine learning. Reducing input data can be done by using only a subset of the original input data. This is called sampling. There are different ways of conducting sampling (e.g., simple random sampling, systematic sampling), As an example, Verdecchia et al. (2022) used stratified sampling, which means randomly selecting data points from homogeneous subgroups of the original dataset.
Participant
Data Scientist
Related software artifact
Data
Context
Machine Learning
Software feature
Data Preprocessing
Tactic intent
Enhance energy efficiency by using a subset of the original input data for training and inference
Target quality attribute
Energy Efficiency
Other related quality attributes
Accuracy, Data Representativeness
Measured impact
Sampling can lead to savings in energy consumption during model training with only negligible reductions in accuracy. Verdecchia et al (2022) achieved decrease in energy consumption of up to 92%