This white paper introduces my educational community initiative to learn how to run AI, ML and other emerging workloads in the most efficient and cost-effective way across diverse models, data sets, software and hardware.
In data-centric AI, active learning (AL) plays a vital role, but current AL tools 1) require users to manually select AL strategies, and 2) can not perform AL tasks efficiently.
Federated learning (FL) is a rapidly growing research field in machine learning.
We present CodeReef - an open platform to share all the components necessary to enable cross-platform MLOps (MLSysOps), i. e. automating the deployment of ML models across diverse systems in the most efficient way.
It improves the training efficiency, remarkably relaxes the requirements on the hardware, and supports efficient large-scale FL experiments with stateful clients by: (1) sequential training clients on devices; (2) decomposing original aggregation into local and global aggregation on devices and server respectively; (3) scheduling tasks to mitigate straggler problems and enhance computing utility; (4) distributed client state manager to support various FL algorithms.
AI engineering has emerged as a crucial discipline to democratize deep neural network (DNN) models among software developers with a diverse background.
This article provides an overview of the Collective Knowledge technology (CK or cKnowledge).
In today's dynamic technological landscape, sustainability has emerged as a pivotal concern, especially with respect to architecting Machine Learning enabled Systems (MLS).
Software Engineering
In recent years, machine learning has transitioned from a field of academic research interest to a field capable of solving real-world business problems.
Traditional data lakes provide critical data infrastructure for analytical workloads by enabling time travel, running SQL queries, ingesting data with ACID transactions, and visualizing petabyte-scale datasets on cloud storage.