Skip to content

13 proper handling of batches #17

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 27 commits into from
Nov 9, 2021
Merged

Conversation

pzimbrod
Copy link
Contributor

@pzimbrod pzimbrod commented Nov 2, 2021

Resolves #13

@pzimbrod
Copy link
Contributor Author

pzimbrod commented Nov 9, 2021

Following things had to be changed:

  • For sake of extensibility, NNlib's batched routines have been ditched by OMEinsum. Looping over 3-dimensional arrays would suffice for a 1D Problem, but certainly not for higher-dimensional ones.
  • Pre-allocation of arrays had to be dropped for now. In combination with OMEinsum, gains were marginal at best anyway.
  • Fixed a problem where the constructor and subsequent training wouldn't work when bias was set to false.
  • Typed every part of the data structure. This should allow the LLVM Compiler to produce more optimized code.
  • Ditched the batchsize argument out of the initialization function as it's no longer needed. This increases the flexibility of the layer considerably.
  • Had to do some permutation at the beginning and end of the layer pass. Otherwise it's not possible to satisfy the sequential requirement of batch dims that CuFFT poses and the requirements of Flux.DataLoader at the same time.

@pzimbrod pzimbrod merged commit cb41f77 into master Nov 9, 2021
@pzimbrod pzimbrod deleted the 13-proper-handling-of-batches branch January 31, 2022 07:20
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Proper handling of batches
1 participant