Closed p3nGu1nZz closed 1 month ago
we completed a majority of this issue, however when we build the token table its not populating. I suspect this is due to the different table sizes of our embedding from 384 to 3. i think the table size is hardcoded in many places which needs investigating.
the _opti file is generating correctly with our pca sklearn algorithm.
Is your feature request related to a problem? Please describe. Currently, our database implementation only supports fixed table sizes (384 columns). We need to import
token_reduce.json
into the database, which requires supporting variable table sizes based on the number of PCA components.Describe the solution you'd like Create a new function in our database code that:
token_reduce.json
into the database.Describe alternatives you've considered
Additional context This change is necessary to handle the reduced embeddings efficiently and flexibly. The new function should be able to:
token_reduce.json
file.Example structure of
token_reduce.json
: