awesome-seml
A curated list of articles that cover the software engineering best practices for building machine learning applications.
https://github.com/SE-ML/awesome-seml
Last synced: 1 day ago
JSON representation
-
Model Training
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- How do you manage your Machine Learning Experiments?
- Machine Learning Testing: Survey, Landscapes and Horizons
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On human intellect and machine failures: Troubleshooting integrative machine learning systems
- Pitfalls and Best Practices in Algorithm Configuration
- Pitfalls of supervised feature selection
- Preparing and Architecting for Machine Learning
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- Machine Learning Testing: Survey, Landscapes and Horizons
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Software development best practices in a deep learning environment
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Pitfalls and Best Practices in Algorithm Configuration
- Software development best practices in a deep learning environment
- What Went Wrong and Why? Diagnosing Situated Interaction Failures in the Wild
- How do you manage your Machine Learning Experiments?
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Fairness On The Ground: Applying Algorithmic FairnessApproaches To Production Systems
- Nitpicking Machine Learning Technical Debt
- Preparing and Architecting for Machine Learning
- Preliminary Systematic Literature Review of Machine Learning System Development Process
- Apples-to-apples in cross-validation studies: pitfalls in classifier performance measurement
- On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach
- Pitfalls of supervised feature selection
- 10 Best Practices for Deep Learning
- On human intellect and machine failures: Troubleshooting integrative machine learning systems
-
Deployment and Operation
- TFX: A tensorflow-based Production-Scale ML Platform
- Versioning for end-to-end machine learning pipelines
- ML Ops: Machine Learning as an engineered disciplined
- Building Continuous Integration Services for Machine Learning
- Machine learning: Moving from experiments to production
- ML Ops: Machine Learning as an engineered disciplined
- ModelOps: Cloud-based lifecycle management for reliable and trusted AI
- Scaling Machine Learning as a Service
- The ML Test Score: A Rubric for ML Production Readiness and Technical Debt Reduction
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- Continuous Training for Production ML in the TensorFlow Extended (TFX) Platform
- Model Governance Reducing the Anarchy of Production
- ModelOps: Cloud-based lifecycle management for reliable and trusted AI
- Operational Machine Learning
- ML Ops: Machine Learning as an engineered disciplined
- Versioning for end-to-end machine learning pipelines
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- Fairness Indicators: Scalable Infrastructure for Fair ML Systems
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- ML Ops: Machine Learning as an engineered disciplined
- Continuous Delivery for Machine Learning
- Machine Learning Logistics
- Scaling Machine Learning as a Service
- ML Ops: Machine Learning as an engineered disciplined
- Best Practices in Machine Learning Infrastructure
- Building Continuous Integration Services for Machine Learning
- Fairness Indicators: Scalable Infrastructure for Fair ML Systems
- Machine learning: Moving from experiments to production
- The ML Test Score: A Rubric for ML Production Readiness and Technical Debt Reduction
- Underspecification Presents Challenges for Credibility in Modern Machine Learning
-
Social Aspects
-
Broad Overviews
- AI Engineering: 11 Foundational Practices
- Best Practices for Machine Learning Applications
- Rules of Machine Learning: Best Practices for ML Engineering
- AI Engineering: 11 Foundational Practices
- Best Practices for Machine Learning Applications
- Hidden Technical Debt in Machine Learning Systems
- Engineering Best Practices for Machine Learning
- Software Engineering for Machine Learning: A Case Study
- AI Engineering: 11 Foundational Practices
-
Data Management
- Automating Large-Scale Data Quality Verification
- How to organize data labelling for ML
- The Data Linter: Lightweight, Automated Sanity Checking for ML Data Sets
- Automating Large-Scale Data Quality Verification
- The Data Linter: Lightweight, Automated Sanity Checking for ML Data Sets
- A Survey on Data Collection for Machine Learning A Big Data - AI Integration Perspective_2019
- How to organize data labelling for ML
- The curse of big data labeling and three ways to solve it
- The ultimate guide to data labeling for ML
- Data management challenges in production machine learning
- Data Validation for Machine Learning
-
Governance
- A Human-Centered Interpretability Framework Based on Weight of Evidence
- Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims
- Closing the AI Accountability Gap: Defining an End-to-End Framework for Internal Algorithmic Auditing
- Inherent trade-offs in the fair determination of risk scores
- Understanding Software-2.0
- A Human-Centered Interpretability Framework Based on Weight of Evidence
- An Architectural Risk Analysis Of Machine Learning Systems
- Beyond Debiasing
- Responsible AI practices
-
Tooling
- Aim - Aim is an open source experiment tracking tool.
- FairLearn - A toolkit to assess and improve the fairness of machine learning models.
- OpenML - An inclusive movement to build an open, organized, online ecosystem for machine learning.
- Alibi Detect - Python library focused on outlier, adversarial and drift detection.
- Tensorflow Data Validation (TFDV) - Library for exploring and validating machine learning data. Similar to Great Expectations, but for Tensorflow data.
- Archai - Neural architecture search.
- Data Version Control (DVC) - DVC is a data and ML experiments management tool.
- Facets Overview / Facets Dive - Robust visualizations to aid in understanding machine learning datasets.
- Great Expectations - Data validation and testing with integration in pipelines.
- HParams - A thoughtful approach to configuration management for machine learning projects.
- Kubeflow - A platform for data scientists who want to build and experiment with ML pipelines.
- Label Studio - A multi-type data labeling and annotation tool with standardized output format.
- Model Card Toolkit - Streamlines and automates the generation of model cards; for model documentation.
- FairLearn - A toolkit to assess and improve the fairness of machine learning models.
- OpenML - An inclusive movement to build an open, organized, online ecosystem for machine learning.
- TensorBoard - TensorFlow's Visualization Toolkit.
- PyTorch Lightning - The lightweight PyTorch wrapper for high-performance AI research. Scale your models, not the boilerplate.
- Robustness Metrics - Lightweight modules to evaluate the robustness of classification models.
- Seldon Core - An MLOps framework to package, deploy, monitor and manage thousands of production machine learning models on Kubernetes.
- Spark Machine Learning - Spark’s ML library consisting of common learning algorithms and utilities.
- Tensorflow Extended (TFX) - An end-to-end platform for deploying production ML pipelines.
- Airflow - Programmatically author, schedule and monitor workflows.
- LiFT - Linkedin fairness toolkit.
- MLFlow - Manage the ML lifecycle, including experimentation, deployment, and a central model registry.
- Neptune.ai - Experiment tracking tool bringing organization and collaboration to data science projects.
- Neuraxle - Sklearn-like framework for hyperparameter tuning and AutoML in deep learning projects.
- REVISE: REvealing VIsual biaSEs - Automatically detect bias in visual data sets.
Programming Languages
Categories
Sub Categories
Keywords
machine-learning
7
deep-learning
6
mlops
3
python
3
data-science
2
hyperparameters
2
pipeline
2
hyperparameter-optimization
2
pytorch
2
data-profiling
1
data-quality
1
data-unit-tests
1
datacleaner
1
datacleaning
1
dataquality
1
dataunittest
1
eda
1
exploratory-analysis
1
exploratory-data-analysis
1
exploratorydataanalysis
1
pipeline-debt
1
pipeline-testing
1
pipeline-tests
1
aiops
1
data-visualization
1
model-cards
1
responsible-ai
1
responsible-ml
1
tensorflow
1
transparency
1
adversarial
1
anomaly
1
concept-drift
1
data-drift
1
detection
1
drift-detection
1
images
1
outlier
1
semi-supervised-learning
1
tabular-data
1
text
1
time-series
1
unsupervised-learning
1
cleandata
1
data-engineering
1
data-profilers
1
python3
1
annotation
1
annotation-tool
1
annotations
1