floydnet 0.1.2__tar.gz → 1.0.0__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: floydnet
3
- Version: 0.1.2
3
+ Version: 1.0.0
4
4
  Summary: Floyd Multi-Head Attention: a drop-in variant of PyTorch MHA with module and function APIs
5
5
  Project-URL: Homepage, https://github.com/ocx-lab/FloydNet
6
6
  Project-URL: Repository, https://github.com/ocx-lab/FloydNet
@@ -235,7 +235,7 @@ Description-Content-Type: text/markdown
235
235
  [![Python](https://img.shields.io/badge/Python-3.9%2B-blue)](https://www.python.org/)
236
236
  [![PyTorch](https://img.shields.io/badge/PyTorch-2.1%2B-orange)](https://pytorch.org/)
237
237
 
238
- Official implementation of an ICLR paper (TODO: add paper title, authors, and links/arXiv).
238
+ Official implementation of [FloydNet](https://openreview.net/pdf?id=aUsx1G6RVQ).
239
239
 
240
240
  ![Figure Pivotal Attention Mechanism for 2-Floyd/3-Floyd.](misc/pivotalattn2&3.png)
241
241
 
@@ -247,7 +247,7 @@ This repository serves two audiences:
247
247
 
248
248
  ## Introduction
249
249
 
250
- FloydNet is the official PyTorch implementation accompanying an ICLR paper (TODO).
250
+ FloydNet is the official PyTorch implementation.
251
251
  The repository provides:
252
252
 
253
253
  1. **Reusable components**: a drop-in attention/Transformer-block interface intended for integration into existing projects.
@@ -374,7 +374,7 @@ If you use this code in your research, please cite the paper:
374
374
  @inproceedings{TODO,
375
375
  title = {TODO},
376
376
  author = {TODO},
377
- booktitle = {International Conference on Learning Representations (ICLR)},
377
+ booktitle = {TODO},
378
378
  year = {TODO},
379
379
  url = {TODO}
380
380
  }
@@ -3,7 +3,7 @@
3
3
  [![Python](https://img.shields.io/badge/Python-3.9%2B-blue)](https://www.python.org/)
4
4
  [![PyTorch](https://img.shields.io/badge/PyTorch-2.1%2B-orange)](https://pytorch.org/)
5
5
 
6
- Official implementation of an ICLR paper (TODO: add paper title, authors, and links/arXiv).
6
+ Official implementation of [FloydNet](https://openreview.net/pdf?id=aUsx1G6RVQ).
7
7
 
8
8
  ![Figure Pivotal Attention Mechanism for 2-Floyd/3-Floyd.](misc/pivotalattn2&3.png)
9
9
 
@@ -15,7 +15,7 @@ This repository serves two audiences:
15
15
 
16
16
  ## Introduction
17
17
 
18
- FloydNet is the official PyTorch implementation accompanying an ICLR paper (TODO).
18
+ FloydNet is the official PyTorch implementation.
19
19
  The repository provides:
20
20
 
21
21
  1. **Reusable components**: a drop-in attention/Transformer-block interface intended for integration into existing projects.
@@ -142,7 +142,7 @@ If you use this code in your research, please cite the paper:
142
142
  @inproceedings{TODO,
143
143
  title = {TODO},
144
144
  author = {TODO},
145
- booktitle = {International Conference on Learning Representations (ICLR)},
145
+ booktitle = {TODO},
146
146
  year = {TODO},
147
147
  url = {TODO}
148
148
  }
@@ -4,7 +4,7 @@ build-backend = "hatchling.build"
4
4
 
5
5
  [project]
6
6
  name = "floydnet"
7
- version = "0.1.2"
7
+ version = "1.0.0"
8
8
  description = "Floyd Multi-Head Attention: a drop-in variant of PyTorch MHA with module and function APIs"
9
9
  readme = "README.md"
10
10
  requires-python = ">=3.9"
File without changes
File without changes
File without changes
File without changes
File without changes