* Add new persistant cache to tune cache
* Serialize autotune presistent cache using vectors
* Properly load and save the persistent cachegf
* Print an error when autotune cache cannot be loaded
* Add tests for persistent cache
Use the same logic as the already implemented tests
* Cargo fmt
* Silence clippy check about implementing default for CliMetricsRenderer
* Add burn-compute feature flag autotune-persistent-cache
This allow burn-compute to remain no-std compliant
* debug
* Git ignore .dir-locals.el files
* Update documentation for compute_checksum implementation
* Expect messages should be an expectation not an error message
* Replace silent eprintln! by log:warn! macro
* Remove clippy allow attribute
* Fix typos in documentation
* Move creation of additional client into the test fn that requires it
* Create compute clients in test function to test different checksum
* Revert tui as a default feature in burn-train cargo file
* Use structs for autotune cache entries
* Unpack InMemoryCacheEntry for even better readibility
* Remove uneeded checksum_checked field in no-std env
* Make sure that autotune cache directoy exists
* Add test for autotune cache file path creation
* Add prefix device info to autotune cache file
* Use new compute in autotune cache integration tests
This prevents race condition by always reloading the cache fir
each test.
* Move burn-compute rand depdencencie in dev-dependencies
* Avoid creation of formatted msg except in case of actual error
* Fix burn-compute unused code warning in no-std env
* Remove the double import for WgpuDevice
* Prioritize wgpu backend over the default ndarray when wgpu feature is set
This fixes `cargo bench --festures wgpu` as `--no-default-features` cannot be
used.
* implement DoubleEndedIterator for DimIter
* return early in next() and next_back()
* return early in next() and next_back()
* remove #[rustfmt::skip]