Lightweight, Portable, Flexible Distributed/Mobile Deep Learning with Dynamic, Mutation-aware Dataflow Dep Scheduler; for Python, R, Julia, Scala, Go, Javascript and more
v2.0.0 Alpha RC3
Adds support for extending MXNet with custom operators, partitioning strategies, and graph passes. All implemented in a library easily compiled separately from the MXNet codebase, and dynamically loaded at runtime into any prebuilt installation of MXNet.
fix for number of inputs/outputs for backward custom ops (#17069) Enhancements for custom subgraph op (#17194) Disable flaky test_custom_op_fork (#17481) fix custom op makefile (#17516) Update CustomOp doc with changes for GPU support (#17486) [WIP] MXNet Extensions enhancements (#17885) (#18128) Dynamic subgraph property (#17034) Dynamic subgraph property doc (#17585) [1.7] Backport MXNet Extension PRs (#17623, #17569, #17762) #18063 (#18069)
[OpPerf] Add Neural network loss ops (#17482) [OpPerf] Fixes the issue when you pass NDArray to run_perf_test (#17508) [OpPerf] Fix markdown for native profile and add profile param in function desc (#17494) [OpPerf] Add Indexing ops (#16253) [OpPerf] Implement remaining random sampling ops (#17502) [OpPerf] Implement remaining GEMM ops (#17501) [OpPerf] Implement all linalg ops (#17528) [OpPerf] Fixed native output ordering, added warmup & runs command line args (#17571) [OpPerf] Add norm, cast ops, remaining optimizer ops (#17542) [Large Tensor] Fixed Embedding op (#17599) [OpPerf] Fixed Python profiler bug (#17642)
Upgrade MKL-DNN dependency to v1.1 (#16823)
Add bfloat16 floating-point format support based on AMP (#17265)
[New Op] Add deformable conv v2 (#16341) Add MXNet Ops for fast multihead attention (#16408) Support boolean elemwise/broadcast binary add, multiply and true_divide (#16728) add gammaln, erf, erfinv (#16811) add aligned roi introduced in Detectron2 (#16619) Implement atleast_1d/2d/3d (#17099) Interleaved MHA for CPU path (#17138) Lamb optimizer update (#16715) Quantized Embedding (#16691) Add gelu fuse ops (#18082) (#18092)
[NumPy] NumPy support for linalg.inv (#16730)
add numpy op nan_to_num (#16717)
[Numpy] Add sampling method for bernoulli (#16638)
Fix numpy-compatible mean output type for integer inputs (#16792)
[Numpy] Fix collect_params().zero_grad() in gluon numpy interface (#16716)
[Numpy][Operator] 'where' Implementation in MXNet (#16829)
[Numpy] Random.normal() with backward (#16330)
Add OP diag [numpy] (#16786)
Mixed precison binary op backward (use in) for numpy (#16791)
add numpy op diagflat [numpy] (#16813)
add op bitwise_or [numpy] (#16801)
[Numpy] Implementation npx.{sample}n (#16876)
[Numpy] Add NumPy support for np.linalg.det and np.linalg.slogdet (#16800)
Op Unravel_index PR [Numpy] (#16862)
[Numpy] Fix imperative basic indexing in numpy (#16902)
[Numpy] Basic indexing in symbolic interface of DeepNumpy (#16621)
[Numpy] add op full_like, c++ impl, fix zeros_like, ones_like type inference (#16804)
[Numpy] Implement numpy operator 'average' (#16720)
[Bugfix] [Numpy] Add kAddTo
and kNullOp to Transpose (#16979)
set rtol = 1e-2 and atol = 1e-4 when dtype == np.float32 in test_numpy_op.py:test_np_linalg_solve (#17025)
Op_Diagonal [Numpy] (#16989)
numpy bincount (#16965)
[numpy] add op bitwise_not (#16947)
[Numpy ]Modify np.random.shuffle to enable inplace by default (#17133)
[numpy] fix argsort typo (#17150)
[numpy] add op round (#17175)
[numpy]Add op delete (#17023)
[numpy] add op flipud, fliplr (#17192)
[CI] Re-enable testing with numpy 1.18 (#17200)
[Numpy] Add broadcast_to scalar case (#17233)
[Numpy] Random.gamma() implemented (#16152)
[Numpy] add row_stack (=vstack) (#17171)
[Numpy] Add infra for performing constraint check (#17272)
porting numpy-compatible hstack to master and add dstack for interoperability (#17030)
adding asnumpy() to output of gather(implicitly called) to fix gather test in large vector and tensor tests (#17290)
[numpy] add op random.exponential (#17280)
[NumPy] Add NumPy support for norm (#17014)
[numpy]add op random.lognormal (#17415)
Add numpy random weibull operator (#17505)
[numpy] Add np.random.pareto and np.random.power (#17517)
[Numpy] Add sort op (#17393)
[numpy]implement exponential backward (#17401)
[Numpy] Where operator scalar version (#17249)
[numpy] add op matmul (#16990)
[numpy]add op random.logistic, random.gumbel (#17302)
[numpy][Do Not Review]add op insert (#16865)
[numpy] add op random.rayleigh (#17541)
[numpy] add fallback ops (#17609)
[numpy] add op pad (#17328)
[numpy] add op fabs, sometrue, round (#17619)
Add arange_like to npx (#16883)
try to move shape_array to npx (#16897)
support np.argsort (#16949)
np.broadcast_to extension (#17358)
support bitwise_and (#16861)
fix np.argmax/argmin output data type (#17476)
add op random.beta (#17390)
add op isnan isinf (#17535)
array_split pr (#17032)
Mixed data type binary ops (#16699)
randn implemented (#17141)
refactor and reduce float types for some functions, also add bitwise_xor (#16827)
any/all (#17087)
amax (#17176)
fix format (#17100)
add op empty_like, add nan_to_num to dispatch (#17169)
handle array_like fill_value for np.full; add unit test coverage (#17245)
add np.amin (#17538)
add npx.gather_nd (#17477)
add np.random.chisquare (#17524)
add polyval (#17416)
add isposinf isneginf isfinite (#17563)
Support broadcast assign for npi_boolean_mask_assign_tensor
(#17131)
Implement Weibull backward (#17590)
support np.dsplit, fix some error msgs and corner cases for hsplit and vsplit, add interoperability tests for h/v/dsplit (#17478)
add np.product (#17489)
Implement np.random.pareto backward (#17607)
add np.ediff1d (#17624)
more support for boolean indexing and assign (#18352)
Fix einsum gradient (#18482)
[v1.7.x] Backport PRs of numpy features (#18653)
[v1.7.x] backport mixed type binary ops to v1.7.x (#18649)
revise activations (#18700)
[Large Tensor] Add support to Random Sample & Pdf ops (#17445) [Large Tensor] Add LT support for NN optimizers and 1 activation function (#17444) [Large Tensor] Fixed SoftmaxActivation op (#17634) [Large Tensor] Fixed col2im op (#17622) [Large Tensor] Fixed Spatial Transformer op (#17617) [Large Tensor] Fix ravel_multi_index op (#17644) Sparse int64 Large tensor support (#16898) Re-Enabling Large Tensor Nightly on GPU (#16164) enabling build stage gpu_int64 to enable large tensor nightly runs (#17546)
MKLDNN FC : Add error info when mkldnn fc bias dimension is wrong (#16692) [MKLDNN] support mkldnn gelu (#16710) [MKLDNN] Fix int8 convolution/fc bias overflow (#16734) [MKLDNN] use dim_t instead of int in slice/transpose operators (#16737) Mkldnn fullyConnect bwd bug fix (#16890) Revert Mkldnn fullyConnect bwd bug fix (#16890) (#16907) [MKLDNN] Use MKLDNNRun (#16772) [MKLDNN] mkldnn RNN operator enhancement (#17075) [MKLDNN] enable MaxPooling with full pooling convention (#16860) update mkldnn to v1.1.2 (#17165) improve mkldnn doc (#17198) [MKLDNN] Fix _copyto (#17173) [MKLDNN] Support channel wise quantization for FullyConnected (#17187) fixed seed for mkldnn test (#17386) add mkldnn softmax backward (#17170) cmake: copy dnnl headers to include/mkldnn (#17647) [mkldnn]Mkldnn bn opt backport from master to 1.7x (#18009) [v1.x] Update 3rdparty/mkldnn remote URL and pin to v1.3 (#17972) (#18033) [v1.x] backport #17900 [MKLDNN] support using any format in pooling backward (#18067) Static link MKL-DNN library (#16731) Add large tensor nightly tests for MKL-DNN operators (#16184) [MKL-DNN] Enable and Optimization for s8 eltwise_add (#16931) [MKL-DNN] Enhance Quantization Method (#17161) Static Build and CD for mxnet-cu102/mxnet-cu102mkl (#17074) MKL-DNN RNN backward path enhancement (#17183) cmake: check USE_OPENMP and pass proper MKL-DNN build flags (#17356) update mkl to 2020.0 (#17355) Enable MKL-DNN by default in pip packages (#16899) Enable MKL-DNN FullyConnected backward (#17318) Softmax primitive cache and in-place computation (#17152) boolean_mask_assign with start_axis (#16886) use identity_with_cast (#16913) change error tolerance for bf16 bn (#18110) [v1.x] Backport #17689 and #17884 to v1.x branch (#18064) refactor codes and add an option to skip/check weight's version to reduce overhead (#17707) (#18039) [v1.x] Backport #17702 and #17872 to v1.x branch (#18038)
Update TensorRT tutorial to build-from-source. (#14860) Minor fix, use RAII for TensorRT builder and network object (#17189)
Add silent option to quantization script (#17094)
Implemented final two binary ops, added default params for functionality (#17407) Implement remaining nn_activation ops in opperf (#17475) Implement all miscellaneous ops (#17511) Implement remaining nn_basic ops in opperf (#17456)
Fix memory leak reported by ASAN in NNVM to ONNX conversion (#15516) ONNX export: Gather (#15995) ONNX export: Slice op - Handle None value for ends (#14942)
[Model] Implement Neural Collaborative Filtering with MXNet (#16689) Further optimization for NCF model (#17148) HMM Model (#17120)
Faster GPU NMS operator (#16542)
[MXNET-1421] Added (CuDNN)BatchNorm operator to the list of mirrored operators (#16022)
dynamic custom operator support (#15921)
Multi Precision Lamb Update operator (#16885)
Add im2col and col2im operator (#16502)
Quantized Elemwise Mul Operator (#17147)
Enhancements for MXTensor for custom operators (#17204)
Enabling large tensor support for binary broadcast operators (#16755)
Fix operators lying about their number of inputs (#17049)
[WIP] Fallback mechanism for mx.np operators (#16923)
Dynamic custom operator GPU support (#17270)
Fix flaky - test_operator_gpu.test_np_insert (#17620)
MXNet FFI for Operator Imperative Invocation (#17510)
[MXNET-978] Higher Order Gradient Support logp1
, expm1
, square
. (#15416)
[MXNET-978] Higher Order Gradient Support arcsin
, arccos
. (#15515)
[MXNET-978] Higher Order Gradient Support rsqrt
, rcbrt
. (#15476)
gather_nd: check bound and wrap negative indices (#17208)
Remove dilation restriction for conv3d (#17491)
Fix storage type infer of softmax backward (#17576)
Fix and optimize handling of vectorized memory accesses (#17767) (#18113)
Cherry-pick of #17995 and #17937 to 1.x branch (#18041)
No tensor cores for fp32 interleaved attention, remove div by 8 restriction (#17994) (#18085)
GPU gemms true fp16 (#17466) (#18023)
Add support for boolean inputs to FusedOp (#16796)
[BUG FIX] Always preserve batch dimension in batches returned from dataloader (#16233)
Fix SliceChannel Type inference (#16748)
change _generate_op_module_signature get_module_file open with encoding=utf-8,it fix some encode error in Chinese windows system. (#16738)
Fix rtrue_divide grad (#16769)
fix inv test flakiness using random matrices generated by SVD (#16782)
[MXNET-1426] Fix the wrong result of sum, mean, argmin, argmax when inputs contain inf or nan (#16234)
Fix (#16781)
fix expand_dims fall back when input's ndim is 0 (#16837)
[fix] missing input log higher order. (#15331)
Fix IndentationError in setup.py (#16857)
Fix a few np issues (#16849)
Fix InferAttr/InferShapeAttr not calling inference for all nodes in a graph (#16836)
fix for enable model parallelism for non-fp32 data (#16683)
Fix NDArrayIter iteration bug when last_batch_handle='pad' (#16166)
Fix crashing on Windows in ObjectPool ~ctor (#16941)
Fix NDArrayIter cant pad when size is large (#17001)
fix axis=-1 bug (#17016)
Fix CUDNN detection for CMake build (#17019)
Fix omp assert issue (#17039)
mshadow: fix vector access (#17021)
[BUGFIX] Fix race condition in kvstore.pushpull (#17007)
[BUGFIX] Fix trainer param order (#17068)
[BugFix] fix filter channel calculation in ModulatedDeformableConvV2 (#17070)
Fix reshape interoperability test (#17155)
fix norm sparse fallback (#17149)
fix py27 quantization (#17153)
fix int8 add ut (#17166)
Fix and clean up Ubuntu build from source instructions (#17229)
fix lstm layer with projection save params (#17266)
Fix rendering of ubuntu_setup.md codeblocks (#17294)
Fix #17267, add expected and got datatype for concat error msgs (#17271)
[BUGFIX] fix model zoo parallel download (#17372)
fix use int8, uint8, int32, int64 (#17188)
[Fix] Add ctx to the original ndarray and revise the usage of context to ctx (#16819)
Fix ndarray indexing bug (#16895)
fix requantize flaky test (#16709)
Initial checkin (#16856)
Fix flakey test_ndarray.py:test_reduce (#17312)
fix flaky test: boolean index and fix bugs (#17222)
Fix IOT Devices section of Get Started page (#17326)
add logic for no batch size while getting data arrays from executors (#17772) (#18122)
Fix reverse shape inference in LayerNorm (#17683)
fix full and full_like when input is boolean (#17668)
Fix MBCC inference (#17660)
Additional fix for vector access. (#17230)
Cherrypick Fix nightly large_vector test caused by incorrect with_seed path (#18178) (#18220)
[1.7] Pass args fix3 (#18237)
fixing batch_norm and layer_norm for large tensors (#17805) (#18261)
[1.7.x] Backport of LSTM and GRU fix (#17898) and RNN op (#17632) (#18316)
[v1.7.x] backport #18500 - [Bug Fixed] Fix batch norm when grad_req is add
(#18517)
Fix the monitor_callback invalid issue during calibration with variable input shapes (#18632) (#18703)
Fix the problem in printing feature in c++ API examples : feature_extract (#15686) updating MXNet version to 1.6.0 in base.h for C APIs (#16905) [API] unified API for custom kvstores (#17010) fix parameter names in the estimator api (#17051) adding docs for 64bit C APIs of large tensor (#17309) Add API docs to INT64 APIs (#16617)
[Quantization] Enhance gluon quantization API (#16695) [Gluon] Improve estimator usability and fix logging logic (#16810) Fix test_gluon.py:test_sync_batchnorm when number of GPUS > 4 (#16834) [Gluon] Update contrib.Estimator LoggingHandler to support logging per batch interval (#16922) Include eval_net the validation model in the gluon estimator api (#16957) Fix Gluon Estimator nightly test (#17042) [MXNET-1431] Multiple channel support in Gluon PReLU (#16262) Fix gluon.Trainer regression if no kvstore is used with sparse gradients (#17199) refactor gluon.utils.split_data() following np.array_split() (#17123) Add RandomApply in gluon's transforms (#17242) Partitioning Gluon HybridBlocks (#15969) Random rotation (#16794) bump up atol for gradient check (#16843) Extend estimator.evaluate() to support event handlers (#16971) [MXNET-1438] Adding SDML loss function (#17298)
Add unoptimized symbol to executor for sharing (#16798) Enforces NDArray type in get_symbol (#16871) Fix #17164 symbolblock with BatchNorm inside during cast to fp16 (#17212) autograd video and image link fixes and removing symbol tutorials (#17227) Fix CosineEmbeddingLoss in when symbol API is used (#17308) Fix Horovod build error due to missing exported symbols (#17348) Update symbol.py (#17408) update symbol to json (#16948)
Python 2 compatibility fix in base.py adding stacktrace in Jenkinsfile_utils.groovy to inspect Python2 failure cause in CI (#17065) Fix image display in python autograd tutorial (#17243) Fix Python 3 compatibility in example/speech_recognition (#17354) Stop testing Python 2 on CI (#15990) Docs: Python tutorials doc fixes (#17435) pin python dependencies (#17556) Python 2 cleanup (#17583)
Simplify C++ flags (#17413)
fix R docs (#16733) [R package] Make R package compilation support opencv 4.0 (#16934) Support R-package with cmake build and fix installation instructions (#17228) Fix R-package/src/Makevars for OpenCV 4 (#17404) Fix typo in Install the MXNet Package for R (#17340)
[MXNET-1440] julia: porting current_context
(#17142)
julia: porting context.empty_cache
(#17172)
pin Markdown version to 3.1 in Julia doc build (#17549)
[Perl] - ndarray operator overloading enhancements (#16779) MXNET-1447 [Perl] Runtime features and large tensor support. (#17610)
Fix scala publish & nvidia-docker cublas issue (#16968) Fix publishing scala gpu with cpu instance (#16987) swap wget to curl in Scala scripts (#17041) [Scala/Java] Remove unnecessary data slicing (#17544) quantile_scalar (#17572) Fix get_started scala gpu (#17434) Fix MBCC & scala publish pipeline (#17643) Bump up additional scala 1.x branch to 1.7.0 (#17765)
Build.py improvement (#16976) Improvements to config.cmake (#17639) [Done] BilinearResize2D optimized (#16292) Speed fused_op compilation by caching ptx and jit-compiled functions (#16783) Improve the speed of the pointwise fusion graph pass (#17114) broadcast_axis optimization (#17091) Optimize AddTakeGrad Tensor Sum (#17906) (#18045)
Add CustomOp tutorial doc (#17241) Correct the grammar in 1-ndarray tutorial (#17513)
Website edits (#17050) [Website 2.0] Nightly Build for v1.x (#17956) [docs] Fix runtime feature detection documentation (#16746) Adding user guidelines for using MXNet built with Large Tensor Support (#16894) fix typo and doc (#16921) large tensor faq doc fix (#16953) [DOC] Add a few tips for running horovod (#17235) Update NOTICE to fix copyright years (#17330) [DOC] Fix tutorial link, and better error msg (#17057) doc fix for argmax & argmin (#17604)
support mixed-precision true_divide (#16711) Try to fix CI (#16908) mixed precision for power (#16859) Fix desired precision for test_ndarray.py:test_reduce (#16992) [reproducibility] multi_sum_sq review, AtomicAdd removal (#17002) fix precision problem in linalg_solve, linalg_tensorinv, linalg_cholesky op test (#16981) grouping large array tests based on type and updating nightly CI function (#17305) [LICENSE] fix cpp predcit license (#17377) [CI] Fix static build pipeline (#17474) skipping tests that cannot fit in nightly CI machine corrected imports (#17450) Update Windows CI scripts to use syntax compatible with Win 2019 server powershell. (#17526) Fix Non-ASCII character in docstring (#17600) [CI] Follow redirects when downloading apache-maven-3.3.9-bin.tar.gz (#17608) [CI] Upgrade sphinx and autodocsumm (#17594) Reduce load on CI due to excessive log flood (#17629) Enable users to specify BLAS (#17648) [CI] Add AMI id to instance info on builds (#17649) [v1.7.x] Backport staggered CI builds (#17999 & #18119) (#18142) [v1.7.x] Backport #17177 to 1.7.x (Fix incorrect calculation results when the C locale is set to a locale that uses commas as the decimal separator) (#18147) Fix formatting and typos in CD README.md (#16703) [CD] dynamic libmxet pipeline fix + small fixes (#16966) [CD] enable s3 publish for nightly builds in cd (#17112) [CD] fix CD pipeline (#17259) [CD] update publish path (#17453) fix CD and remove leftover from #15990 (#17551) Fix nightly build (#16773) Update pypi_publish.py to disable nighlty build upload to Pypi (#17082) [v1.7.x] update jetson dockerfile to support CUDA 10.0 (#18339) Remove manually created symbolic link to ninja-build (#18437) (#18456) Increase staggered build timeout to 180 min (#18568) (#18585)
Don't relicense FindCUDAToolkit.cmake (#17334) fix license and copyright issues (#17364) Update ps-lite LICENSE (#17351) remove unused file with license issue (#17371) Update LICENSE for fonts (#17365) license np_einsum file under bsd (#17367) Update Apache License for mshadow (#18109) (#18134)
Link fixes4 (#16764) Refactoring names for mxnet version of nnvm to avoid conflicting with the original tvm/nnvm. (#15303) minor typo fix (#17008) Add micro averaging strategy to pearsonr metric (#16878) introduce gradient update handler to the base estimator (#16900) fix latency calculation and print issue (#17217) add inference benchmark script (#16978) change the wording and log level to be more in line with the general use (#16626) Updated logos. (#16719) Pinning rvm version to satisfy Jekyll build (#18016) Workaround gnu_tls handshake error on Ubuntu 14.04 Nvidia Docker (#18044)
Please follow the instructions at https://mxnet.incubator.apache.org/get_started
name | commit-id | last updated in MXNet | last update in module |
---|---|---|---|
dlpack | 3efc489 | Jan 20, 2020 | Feb 16, 2020 |
dmlc-core | b3a4c71 | Dec 10, 2019 | Apr 25, 2020 |
googletest | eb9225c | Jan 14, 2019 | Apr 16, 2020 |
mkldnn | 07579e6 | Mar 31, 2020 | Apr 24, 2020 |
nvidia_cub | c3cceac | Feb 16, 2018 | Jul 17, 2019 |
onnx-tensorrt | f4745fc | Jul 12, 2019 | Apr 23, 2020 |
openmp | b76842e | Jul 18, 2019 | Oct 15, 2019 |
ps-lite | f601054 | Jan 24, 2020 | Feb 28, 2020 |
tvm | 9bd2c7b | Jan 23, 2020 | Apr 26, 2020 |
WARNING: THIS IS NOT AN APACHE SOFTWARE FOUNDATION RELEASE OF MXNET AS IT PREDATES MXNET JOINING THE APACHE SOFTWARE FOUNDATION
WARNING: THIS IS NOT AN APACHE SOFTWARE FOUNDATION RELEASE OF MXNET AS IT PREDATES MXNET JOINING THE APACHE SOFTWARE FOUNDATION
MXNet community voted to no longer support Python 2 in future releases of MXNet. Therefore, MXNet 1.6 release is going to be the last MXNet release to support Python 2.
NumPy has long been established as the standard math library in Python, the most prevalent language for the deep learning community. With this library as the cornerstone, there are now the largest ecosystem and community for scientific computing. The popularity of NumPy comes from its flexibility and generality.
In #14253, the MXNet community reached consensus on moving towards a NumPy-compatible programing experience and committed to a major endeavor on providing NumPy compatible operators.
The primary goal of the projects below is to provide the equivalent usability and expressiveness of NumPy in MXNet to facilitate Deep Learning model development, which not only helps existing deep learning practitioners but also provides people in the existing NumPy community with a shortcut for getting started in Deep Learning. The efforts towards this goal would also help a secondary goal, which is to enable the existing NumPy ecosystem to utilize GPUs and accelerators to speed up large scale computation.
lcm
, tril
, identity
and take
(#16264)may_share_memory
and shares_memory
(#16533)DL models, besides compute intensive operations like convolutions and fully connected layers, feature a lot of simple pointwise (aka elementwise) operations (like elementwise addition etc.). Performance of those operations is fully memory bandwidth bound and so limit speedups from newer GPU hardware, which typically has high compute/memory bandwidth ratio. When multiple of such operations are chained one after another, it results in a series of unnecessary stores and loads as well as potential increased memory usage to store the intermediate results. Pointwise fusion helps in alleviating those problems by just-in-time generation of fused operators, which do not store intermediate results in memory, resulting in performance and memory usage improvements.
reciprocal
, abs
. (#15413)tan
, tanh
(#15253)arctan
, arctanh
, radians
. (#15531)sqrt
, cbrt
. (#15474)clip
, dropout
. (#15746)sinh
, cosh
. (#15412)arcsinh
, arccosh
. (#15530)self
in warning. (#15614)test_convolution_independent_gradients
(#15939)transpose
(#15865){adam/ftrl/rmprop/rmspropalex}_update
. (#15768)abs
to NDArray and Symbol. (#15680)MXNet community voted to no longer support Python 2 in future releases of MXNet. Therefore, MXNet 1.6 release is going to be the last MXNet release to support Python 2.
MXNET_HOME
to MXNET_ROOT
(#15568)MXNET_HOME
to MXNET_ROOT
(#15568)" (#16147)mx.forward
kwargs checking (#16138)context.num_gpus
(#16236)AbstractMXError
as parent type (#16235)Please follow the instructions at https://mxnet.incubator.apache.org/get_started
Users that build MXNet from source are recommended to build release 1.6.0 without jemalloc to avoid incompatibilities with llvm's openmp library (details in issue #17043 and PR #17324). This is done for cmake builds by setting USE_JEMALLOC "OFF" in ./CMakeLists.txt, or for make builds with "USE_JEMALLOC = 0" in make/config.mk.
Apache MXNet (incubating) 1.5.1 is a maintenance release incorporating important bug fixes and important performance improvements. All users of Apache MXNet (incubating) 1.5.0 are advised to upgrade. You can install Apache MXNet (incubating) 1.5.1 at the usual place. Please review these Release Notes to learn the bug fixes.
MXEnginePushAsyncND
and MXEnginePushSyncND
(#15751) (#15792)Please follow the instructions at https://mxnet.incubator.apache.org/get_started
Name | Commit-id | Last update in MXNet | Last update in module |
---|---|---|---|
dlpack | 10892ac | Oct 30, 2017 | Aug 12, 2019 |
dmlc-core | 3943914 | May 14, 2019 | Sep 2, 2019 |
googletest | eb9225c | Jan 14, 2019 | Aug 29, 2019 |
mkldnn | 41bee20 | May 14, 2019 | Aug 27, 2019 |
mshadow | 1d79ecf | May 13, 2019 | Aug 4, 2019 |
nvidia_cub | c3cceac | Feb 16, 2018 | Jul 17, 2019 |
onnx-tensorrt | 1e209e5 | Jan 3, 2019 | Aug 22, 2019 |
openmp | 37c7212 | Nov 14, 2017 | Aug 28, 2019 |
ps-lite | 8a76389 | Apr 25, 2018 | Sep 2, 2019 |
tvm | 21935dc | May 21, 2019 | Sep 2, 2019 |