frictionlessdata / tableschema-sql-py

Generate SQL tables, load and extract data, based on JSON Table Schema descriptors.
MIT License
60 stars 19 forks source link

Make update work with long scalar values #74

Closed akariv closed 6 years ago

akariv commented 6 years ago

When building the bloom filter, we used to use the results coming from sqlalchemy directly. Turns out that these objects are of type sqlalchemy.engine.result.RowProxy, which for long values truncate the data inside them when converted to string using str (which is exactly what pybloom does before hashing). This causes wrong data to be stored in the bloom filter, causing inserts for existing rows.