Update README.md
Browse files
README.md
CHANGED
|
@@ -4,6 +4,23 @@ datasets:
|
|
| 4 |
- AbstractPhil/geometric-vocab
|
| 5 |
pipeline_tag: zero-shot-classification
|
| 6 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
# After a big notebook refactor
|
| 8 |
|
| 9 |
I have pushed the updated model code, and included the loader. I will not include the losses or the training methodology until the full process is prepared and the paper published. After which you will see exactly what I've developed and why each piece exists. Until then there are only breadcrumbs and inference code.
|
|
|
|
| 4 |
- AbstractPhil/geometric-vocab
|
| 5 |
pipeline_tag: zero-shot-classification
|
| 6 |
---
|
| 7 |
+
# Breakdown and assessment
|
| 8 |
+
|
| 9 |
+
Using the standard vit position tokenization doesn't work. It caps our unique feature map representation to 65 tokens worth based on patch4, and due to the high-dimensional geometry bloating dimensions to such a degree - the representative patches can only be learned to a certain degree of accuracy before they simply run out of space and the model defaults to memorizing the training data.
|
| 10 |
+
|
| 11 |
+
Probe diagnostics show;
|
| 12 |
+

|
| 13 |
+
|
| 14 |
+
They are in fact forming unique separated clusters of similar assessment, and the 2d graph shows they are highly diverse;
|
| 15 |
+

|
| 16 |
+
|
| 17 |
+
So the process works, but the process simply does not have enough compartmentalization to fully represent a classified patch within a classification within the current set of parameters.
|
| 18 |
+
|
| 19 |
+
I am currently devising a better and more directly intertwined structure that represents the baseline of geometry directly entangled with the vit patches rather than trying to shape it using another system indirectly.
|
| 20 |
+
|
| 21 |
+
This direct representation will be a bit more volatile at first but it should help solidify the missing 40% accuracy in a utilizable and repeatable way to extract the necessary patches.
|
| 22 |
+
|
| 23 |
+
|
| 24 |
# After a big notebook refactor
|
| 25 |
|
| 26 |
I have pushed the updated model code, and included the loader. I will not include the losses or the training methodology until the full process is prepared and the paper published. After which you will see exactly what I've developed and why each piece exists. Until then there are only breadcrumbs and inference code.
|