* Add int_random to int tensor ops
* Int random for tch backend
* Int random for burn-fusion
* int random for autodiff
* Int random for candle backend
* Int random for ndarray backend
* Int random for wgpu backend
* Merge imports
* Typo
* Shader file for int uniform distribution
* Create AutotuneOperationSet and public int_sum_dim_autotune
* Adjust bounds to 0..10
* Create uniform_int_kernel, unit tests, use new kernel
* Reduction kernels for regular and shared memory sum_dim int operations
* Macro that accomadates wgpu IntElement
* Add autotuning to int_mean_dim
* Use correct macro for Int autotuning
* Add int_mean_dim_shared_memory
* Add int_mean_dim and unit test
* Create autotunables for mean_dim
* Run fmt
* Remove comment
* Finish resolving merge conflict, fix doc
* Make the element trait bound a parameter to reduce_tune_ops macro
* Update book
* Fix requested change
* Change range to [0, 255] and update test accordingly
* Forgot to include candle in last commit
* Fix comment
* Use correct int autotune for mean dim
* Fix typo- not sure how this passed earlier
* Resolve syntax issues from merge
* Fix cast_float
* Saving here
* Continue fixing merge conflicts, all tests pass locally
* Run fmt
* Change cast_float to cast_u32_to_float
* Make uniform_int_inner_loop safer
* Be even more explicit about u32 casts
* Skip an intermediate step and cast directly to u32
* Replace JitElement + Element with IntElement
* Run fmt
* This should fix the CI
* This time for sure
* Running into issues with identity nodes
* Vec<RefCell<Node>> seems to work for this
* back to passing tests
* Reworked IO into separate struct
* working towards exploiting topological ordering and more informative ident errors
* the passing of an initializer to coalesce is temporary
* cleaning up dead code
* handled unsqueeze
* reworked node initialization and dim inference
* mainly cleanup
* changed how io use is tracked, moved unsqueeze remapping out of dim inference
* `cargo xtask run-checks all` now passes
* added a fixme and a few doc strings
* removing println and dead code
* spaces in doc strings
* altered top sort to work on node proto, moved prior to node gen
* Update ir.rs
* Update from_onnx.rs
removed dead code
* updated doc string
* camalcased Onnx Graph Builder
* removed self import?
* add any, all op implementation for all tensor types
* add op to burn-book
* fix formatting
* refactor tensor operations from numeric to BaseOps.
* fix book doc
* comments fix and add more tests
* PyTorch config deserializer from .pt file
* Update pytorch-model.md
* Format the book section
* Update Cargo.lock
* Recommend to resave config as json
* Fix comment wording
* refactor the publish.rs file in xtask
* fixes after running checks
* remove extra space
* Update messages to be more conventional and follow rust guidelines
---------
Co-authored-by: iamricks <richardsantos@Richards-MacBook-Pro.local>
* fix(book): add missing second parameter to CrosEntropyLoss constructor
CrossEntropyLoss::new() expects two parameters, the pad_index and the device
* fix: fix missing closing parenthese
* [backend-comparison] Add auth command to burnbench CLI
* [backend-comparison] Add --share argument to Burnbench CLI
* Cargo clippy fixes
* Fix typos
* Add comment to explain the FIVE_SECONDS constant
* Use num_args to force at least one arg value and make args required
In the run command, makes the --benches and --backends required
The manual check is no longer necessary
* Use and_then instead of match
* Simplify token verification
* Use map_or instead of match
Combined PRs ➡️📦⬅️✅ The following pull requests have been successfully combined on this PR:
Closes Bump thiserror from 1.0.56 to 1.0.57 #1293 Bump thiserror from 1.0.56 to 1.0.57
Closes Bump tokenizers from 0.15.1 to 0.15.2 #1292 Bump tokenizers from 0.15.1 to 0.15.2
Closes Bump bytemuck from 1.14.1 to 1.14.3 #1291 Bump bytemuck from 1.14.1 to 1.14.3
Closes Bump indicatif from 0.17.7 to 0.17.8 #1290 Bump indicatif from 0.17.7 to 0.17.8