alan-turing-institute/MLJ.jl: v0.10.0
Authors/Creators
- Anthony Blaom, PhD1
- Thibaut Lienart2
- Yiannis Simillides
- Diego Arenas
- vollmersj
- Mosè Giordano3
- Ayush Shridhar4
- Ayush Shridhar
- Okon Samuel
- Ed
- swenkel
- Julian Samaroo
- evalparse5
- Júlio Hoffimann6
- sjvollmer
- Michael Krabbe Borregaard7
- Kevin Squire8
- pshashk
- lhnguyen-vn
- azev779
- Robert Hönig10
- Nils11
- Kryohi
- Julia TagBot
- Evelina Gabasova12
- Dilum Aluthge13
- Cédric St-Jean14
- 1. NeSI/Alan Turing Institute
- 2. AWS
- 3. @UCL-RITS
- 4. IIIT Bhubaneswar
- 5. @evalparse
- 6. IBM Research
- 7. GLOBE Institute
- 8. SecondSpectrum
- 9. Zicklin School of Business, Baruch College
- 10. University of Cambridge
- 11. Queen Mary, University of London
- 12. The Alan Turing Institute
- 13. Brown University
- 14. r2.ca
Description
Upgrade to MLJBase 0.12.0 and MLJModels 0.9.0 to effect the following changes:
[x] (breaking) suppress normalisation of measure weights (MLJBase PR #208)
[x] (breaking) Shift the optional
rngargument of iterator to first position (MLJBase #215)[x] (mildly breaking) Let all models (supervised and unsupervised) share a common set of traits. So, for example, unsupervised models now have the
target_scitypetrait (usually taking the valueUnknown). For a list of the common traits, domodels()[1] |> keys |> collect(https://github.com/alan-turing-institute/MLJBase.jl/issues/163).[x] (enhancement) Add
samplerwrapper for one-dimensional ranges, for random sampling from ranges usingrand(MLJBase #213)[x] Change default value of
num_roundin XGBoost models from 1 to 100 (MLJModels PR #201)
Closed issues:
- Help with loading code on multiple processes for paralleled tuning of a pipeline (#440)
- Re-export CPU1, CPUProcesses, CPUThreads (#447)
- Taking loss functions seriously (#450)
- @pipeline to accept multiple Supervised models (#455)
- What parts of MLJBase should be reexported in MLJ (#462)
unpacknot working (#465)- Automatic Ensembling options (#466)
Merged pull requests:
- Fix model table link in README.md (#467) (@pshashk)
- For a 0.10.0 release (#471) (@ablaom)
Files
alan-turing-institute/MLJ.jl-v0.10.0.zip
Files
(3.5 MB)
| Name | Size | Download all |
|---|---|---|
|
md5:6dbddd361f5efe977b322bd8bfca21de
|
3.5 MB | Preview Download |
Additional details
Related works
- Is supplement to
- https://github.com/alan-turing-institute/MLJ.jl/tree/v0.10.0 (URL)