-
Hi, I am use the create_dvm_dataset.ipynb to generate the required data. I found one problem:
There is no saved file named"dvm_features_{split}_noOH{v}_physical_jittered_50.csv", but the program rea…
-
### Describe your problem in detail.
ATM specification for Tabular files https://bids-specification.readthedocs.io/en/stable/common-principles.html#tabular-files states
> String values containing ta…
-
According to [Tabular files](https://bids-specification.readthedocs.io/en/stable/02-common-principles.html#tabular-files):
> TSV files MUST be in UTF-8 encoding.
We currently don't validate this…
-
The software package is dealing currently with tabular data only. However, there is one important aspect that has not been dealt with categorical variables.
To improve this:
- We need to add detec…
-
Criar ferramenta de dicionário semântico das colunas em diversos sistemas.
Usar método de construção de datapackage em https://github.com/splor-mg/ppag-planejamento-dados-2024/blob/main/scripts/bu…
-
Just shifted from that original Anchor implementation (implemented by paper's author), encounted a problem so wonder if it is exist in this implementation too.
My features had been encoded into one…
-
We're looking for the best way to encode tabular data in json that normally would be stored in a csv, I don't think that dataset is designed to actually store the raw data, but instead it describes a …
-
At the moment the `datapackage-validation` tool require to put the `schema` information within the `resource > vector > schema`. It would be more similar to the `tabular-datapackage` if we put the `sc…
zarch updated
6 years ago
-
Categorical encodings, like one-hot, frequency, and target encoding, are widely used in tabular data.
Does this repo support the categorical features?
-
The tabular data importer ought to read BOM (byte-order-marker). Currently is guessing the encoding line-by-line, which is inefficient and also fails for UTF-16 data (and presumably also UTF-32).