This is the implementation of GPT2 paper in PyTorch. This is just the architecture. It has 162262609 (162M) parameters (not training on my GPU so doesn't matter:) ).
-
Notifications
You must be signed in to change notification settings - Fork 0
yashghogre/GPT2-Implementation
About
This is the implementation of GPT2 architecture in PyTorch.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published