Collective Knowledge (software)

From Wikipedia, the free encyclopedia
Collective Knowledge (CK)
Developer(s)Grigori Fursin and the cTuning foundation
Initial release2015; 9 years ago (2015)
Stable release
2.6.3 (discontinued for the new Collective Mind framework[1]) / November 30, 2022 (2022-11-30)
Written inPython
Operating systemLinux, Mac OS X, Microsoft Windows, Android
TypeKnowledge management, FAIR data, MLOps, Data management, Artifact Evaluation, Package management system, Scientific workflow system, DevOps, Continuous integration, Reproducibility
LicenseApache License for version 2.0 and BSD License 3-clause for version 1.0
Websitegithub.com/ctuning/ck, cknow.io

The Collective Knowledge (CK) project is an open-source framework and repository to enable collaborative, reproducible and sustainable research and development of complex computational systems.[2] CK is a small, portable, customizable and decentralized infrastructure helping researchers and practitioners:

Notable usages[edit]

Portable package manager for portable workflows[edit]

CK has an integrated cross-platform package manager with Python scripts, JSON API and JSON meta-description to automatically rebuild software environment on a user machine required to run a given research workflow.[17]

Reproducibility of experiments[edit]

CK enables reproducibility of experimental results via community involvement similar to Wikipedia and physics. Whenever a new workflow with all components is shared via GitHub, anyone can try it on a different machine, with different environment and using slightly different choices (compilers, libraries, data sets). Whenever an unexpected or wrong behavior is encountered, the community explains it, fixes components and shares them back as described in.[4]

References[edit]

  1. ^ CK package at PYPI
  2. ^ a b Fursin, Grigori (29 March 2021). Collective Knowledge: organizing research projects as a database of reusable components and portable workflows with common APIs. Philosophical Transactions of the Royal Society. arXiv:2011.01149. doi:10.1098/rsta.2020.0211.
  3. ^ Reusable CK components and actions to automate common research tasks
  4. ^ a b c Live paper with reproducible experiments to enable collaborative research into multi-objective autotuning and machine learning techniques
  5. ^ Online repository with reproduced results
  6. ^ Index of reproduced papers
  7. ^ Ed Plowman; Grigori Fursin, ARM TechCon'16 presentation "Know Your Workloads: Design more efficient systems!"
  8. ^ Artifact Evaluation for systems and machine learning conferences
  9. ^ ACM TechTalk about reproducing 150 research papers and testing them in the real world
  10. ^ EU TETRACOM project to combine CK and CLSmith (PDF), archived from the original (PDF) on 2017-03-05, retrieved 2016-09-15
  11. ^ Artifact Evaluation Reproduction for "Software Prefetching for Indirect Memory Accesses", CGO 2017, using CK, 16 October 2022
  12. ^ GitHub development website for CK-powered Caffe, 11 October 2022
  13. ^ Open-source Android application to let the community participate in collaborative benchmarking and optimization of various DNN libraries and models
  14. ^ Reproducing quantum results from nature – how hard could it be?
  15. ^ MLPerf crowd-benchmarking
  16. ^ MLPerf inference benchmark automation guide, 17 October 2022
  17. ^ List of shared CK packages

External links[edit]

  • Development site: [1]
  • Documentation: [2]
  • Public repository with crowdsourced experiments: [3]
  • International Workshop on Adaptive Self-tuning Computing System (ADAPT) uses CK to enable public reviewing of publications and artifacts via Reddit: [4]