Read performance of chunked datasets encoded with the extensible array index as well as the btree2 index has been improved.
Breaking changes
To solve a problem with ambiguity, the signature of the NativeDataset.Read(...) overloads has been changed to make the provision of the datasetAccess parameter mandatory.
v1.0.0-beta.27
Features
Added soft link support
Performance
The lookup performance of chunks in the chunk cache has been improved.
Read performance of chunked datasets encoded with the fixed array index has been improved.
v1.0.0-beta.26
Read performance of multidimensional, contiguous, compact or single-chunked datasets has been dramatically improved by adding special case handling which allows bulk copying of data if the whole dataset is to be read (no slicing).
v1.0.0-beta.25
Read performance of chunked datasets in the old version 1 B-tree format has been dramatically improved by adding a cache to avoid repeated deserialisation of in-file structures.
v1.0.0-beta.24
Bugs fixed
Created a workaround to not throw error "Byte order conversion is not (yet) support by PureHDF." (#101).
v1.0.0-beta.23
Bugs fixed
Fixed a bug where objects were encoded more than once (by using object references) and added a circular reference detection.
v1.0.0-beta.22
Features
Added write support for object references. Example:
var dataset = new H5Dataset(data: 1);
var group = new H5Group();
var file = new H5File
{
["data"] = dataset,
["group"] = group,
["references"] = new H5ObjectReference[] { dataset, group }
};
Read performance of chunked datasets encoded with the extensible array index as well as the btree2 index has been improved.
Breaking changes
To solve a problem with ambiguity, the signature of the NativeDataset.Read(...) overloads has been changed to make the provision of the datasetAccess parameter mandatory.
v1.0.0-beta.27 - 2024-06-21
Features
Added soft link support
Performance
The lookup performance of chunks in the chunk cache has been improved.
Read performance of chunked datasets encoded with the fixed array index has been improved.
v1.0.0-beta.25 - 2024-06-13
Read performance of chunked datasets in the old version 1 B-tree format has been dramatically improved by adding a cache to avoid repeated deserialisation of in-file structures.
v1.0.0-beta.24 - 2024-06-13
Bugs fixed
Created a workaround to not throw error "Byte order conversion is not (yet) support by PureHDF." (#101).
v1.0.0-beta.23 - 2024-05-30
Bugs fixed
Fixed a bug where objects were encoded more than once (by using object references) and added a circular reference detection.
v1.0.0-beta.22 - 2024-05-29
Features
Added write support for object references. Example:
var dataset = new H5Dataset(data: 1);
var group = new H5Group();
var file = new H5File
{
["data"] = dataset,
["group"] = group,
["references"] = new H5ObjectReference[] { dataset, group }
};
Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR:
- `@dependabot rebase` will rebase this PR
- `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it
- `@dependabot merge` will merge this PR after your CI passes on it
- `@dependabot squash and merge` will squash and merge this PR after your CI passes on it
- `@dependabot cancel merge` will cancel a previously requested merge and block automerging
- `@dependabot reopen` will reopen this PR if it is closed
- `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually
- `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency
- `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)
- `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)
- `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
Bumps PureHDF from 1.0.0-beta.18 to 1.0.0.
Release notes
Sourced from PureHDF's releases.
... (truncated)
Changelog
Sourced from PureHDF's changelog.
... (truncated)
Commits
9f989d8
Merge commit 'd0c4806c4cfb32e5af1155994682259022311395'd0c4806
Fix changelog840fad0
Prepare release4901f95
Fix #72: Document how to provide filter options (e.g. Blosc compressor)ae19254
Fix #93: Improve docs: ReadRaw; see also CanRead_Raw_ByteArray, CanRead_Raw_M...023d9e4
Fix #115: Documentation: How to use the chunk cache; add a "performance" section0fb3e31
Fix #85: Honor H5Name for Compound Writingbd89e64
Fix #98: Improve docs: new H5ObjectReference() + implicit conversion123f30e
Fix #92: Improve docs: How to write data with nullable value types?4c605f5
Clean up & fix #112: Add documentation about how to create soft linksDependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting
@dependabot rebase
.Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show