nicelove666 / Allocator-Pathway-IPFSTT

0 stars 0 forks source link

[DataCap Application] Public Datasets -- Digital Earth Africa #3

Closed MaleiWong closed 3 months ago

MaleiWong commented 3 months ago

Version

1

DataCap Applicant

DataFortress

Project ID

001

Data Owner Name

Environment

Data Owner Country/Region

South Africa

Data Owner Industry

Environment

Website

https://www.digitalearthafrica.org/

Social Media Handle

Twitter: https://twitter.com/DEarthAfrica Youtube: https://www.youtube.com/channel/UCBasD3Dz-TdQTRoyG30kVVw

Social Media Type

Slack

What is your role related to the dataset

Data Preparer

Total amount of DataCap being requested

10

Unit for total amount of DataCap being requested

PiB

Expected size of single dataset (one copy)

1

Unit for expected size of single dataset

PiB

Number of replicas to store

10

Weekly allocation of DataCap requested

1

Unit for weekly allocation of DataCap requested

PiB

On-chain address for first allocation

f1bd4ibu3pugyrn6gaxw2t6ebaskr7oj44l6bvzky

Data Type of Application

Public, Open Dataset (Research/Non-Profit)

Custom multisig

Identifier

No response

Share a brief history of your project and organization

DataFortress boasts a team of 20 technical experts who are dedicated to exploring the IPFS and Filecoin domains and possess extensive experience in distributed storage and computing. Our team has a deep understanding of the key technologies of IPFS and Filecoin and has accumulated a wealth of experience in the field of DC data downloading and distribution. We are capable of providing comprehensive technical solutions to clients.

Is this project associated with other projects/ecosystem stakeholders?

No

If answered yes, what are the other projects/ecosystem stakeholders

No response

Describe the data being stored onto Filecoin

Landsat scenes and metadata. DE Africa provides data from Landsat 5, 7 and 8 satellites, including historical observations dating back to late 1980s and regularly updated new acquisitions.

Where was the data currently stored in this dataset sourced from

AWS Cloud

If you answered "Other" in the previous question, enter the details here

No response

If you are a data preparer. What is your location (Country/Region)

Hong Kong

If you are a data preparer, how will the data be prepared? Please include tooling used and technical details?

No response

If you are not preparing the data, who will prepare the data? (Provide name and business)

No response

Has this dataset been stored on the Filecoin network before? If so, please explain and make the case why you would like to store this dataset again to the network. Provide details on preparation and/or SP distribution.

No response

Please share a sample of the data

s3://deafrica-landsat/
s3://deafrica-landsat-inventory/

Confirm that this is a public dataset that can be retrieved by anyone on the Network

If you chose not to confirm, what was the reason

No response

What is the expected retrieval frequency for this data

Yearly

For how long do you plan to keep this dataset stored on Filecoin

1.5 to 2 years

In which geographies do you plan on making storage deals

Greater China, Asia other than Greater China

How will you be distributing your data to storage providers

Cloud storage (i.e. S3), HTTP or FTP server, Shipping hard drives

How did you find your storage providers

Big Data Exchange, Partners

If you answered "Others" in the previous question, what is the tool or platform you used

No response

Please list the provider IDs and location of the storage providers you will be working with.

1.f01025366, ShanDong-QingDao
2.f02940074, JiangXi-FuZhou
3.f02951064, ZheJiang-JinHua
4.f02984282, HeNan-ZhengZhou

How do you plan to make deals to your storage providers

Boost client

If you answered "Others/custom tool" in the previous question, enter the details here

No response

Can you confirm that you will follow the Fil+ guideline

Yes

datacap-bot[bot] commented 3 months ago

Application is waiting for governance review

nicelove666 commented 3 months ago

@MaleiWong I am very happy to see your application. Please answer these questions: If you are a data preparer, how will the data be prepared? Please include tooling used and technical details?

MaleiWong commented 3 months ago

@MaleiWong我很高兴看到你的申请。请回答以下问题: 如果您是数据准备者,将如何准备数据?请包括使用的工具和技术细节?

We carry out secondary development based on https://github.com/karust/gogetcrawl. After downloading a batch of data from Amazon Web Services S3, it will automatically process the packaging and generate the car file. As long as the same download parameters are set, it can be downloaded and Generate car files with the same piececid to ensure that different SPs can download and encapsulate the same file to ensure that the data backup meets the requirements. After the car file is generated, a corresponding unit data will be generated at the same time, including filename, data_cid, piece_cid, piece_size and other information and written into the csv file. The csv file is the datacap distribution file that will be used later, and the file name is the downloaded The marker name and starting position of the dataset. Based on the metadata file, you can tell which portion of the total data set is saved by each LDN. The data download can be executed directly on the SP's device, or we can mail it to the SP through the hard disk after processing.

datacap-bot[bot] commented 3 months ago

Datacap Request Trigger

Total DataCap requested

10PiB

Expected weekly DataCap usage rate

1PiB

Client address

f1bd4ibu3pugyrn6gaxw2t6ebaskr7oj44l6bvzky

datacap-bot[bot] commented 3 months ago

Application is ready to sign

datacap-bot[bot] commented 3 months ago

Request Approved

Your Datacap Allocation Request has been approved by the Notary

Message sent to Filecoin Network

bafy2bzacecu5dhudxxrdlh5ynnw36jijpjsgxg2wy3mrhmfx22rhflflzellu

Address

f1bd4ibu3pugyrn6gaxw2t6ebaskr7oj44l6bvzky

Datacap Allocated

1PiB

Signer Address

f16hmuu3w247dkkhsrbbcbeqbugmpjbxpkrpcdatq

Id

46aa3b66-3253-43f9-bc5a-8aae89f4402e

You can check the status of the message here: https://filfox.info/en/message/bafy2bzacecu5dhudxxrdlh5ynnw36jijpjsgxg2wy3mrhmfx22rhflflzellu

datacap-bot[bot] commented 3 months ago

Application is Granted

nicelove666 commented 3 months ago

I will keep an eye on your distribution plan, please stay compliant

MaleiWong commented 3 months ago

1.f01025366, ShanDong-QingDao 2.f02940074, JiangXi-FuZhou 3.f02951064, ZheJiang-JinHua 4.f02984282, HeNan-ZhengZhou 5.f0122215, ShanDong-QingDao

  1. f0427989, ShanDong-QingDao
  2. f02942808, GuangDong-ShenZhen
  3. f02894875, JiangXi-FuZhou
  4. f02370792, GuangDong-DongGuan We have currently identified the above SPs. Since the total data size and packaging speed that each SP is prepared to store are different, in order to ensure that there are enough copies of the data, we will contact new SPs to participate in the future.
nicelove666 commented 3 months ago

@MaleiWong The amount you apply for every week is too large. Why do you apply for 1PIB? Do you have data processing experience? Can you comply with the rules of FIL+? How to ensure that DC will not be abused?

MaleiWong commented 3 months ago

@MaleiWong The amount you apply for every week is too large. Why do you apply for 1PIB? Do you have data processing experience? Can you comply with the rules of FIL+? How to ensure that DC will not be abused?

DataFortress boasts a team of 20 technical experts who are dedicated to exploring the IPFS and Filecoin domains and possess extensive experience in distributed storage and computing. 2 years of experience in DC data processing.

In v3.1 we wrote weekly requests as 1PiB. But there was only 512TiB in the first round. Therefore, this time we also write the weekly application as 1PiB. We will adjust weekly applications to 200-300TiB in the future.

Compliance is very important. The core of DC is to bring precious data storage assets to Filecoin, which is what we have been pursuing. We have written down our work process in detail in the application form, and you can also download our data to view at any time. We hope to be treated with kindness, thank you.

nicelove666 commented 3 months ago

Sorry, the quota of 1Pib in the first round is still too much. Although I understand your statement, you think that writing 1Pib in the first round may only be 512Tib, so the first round you want to get is 512T.

I suggest you close this ldn and reapply. The weekly application is recommended not to exceed 300T.

datacap-bot[bot] commented 3 months ago

Application is in Refill

datacap-bot[bot] commented 3 months ago

DataCap Allocation requested

Multisig Notary address

Client address

f1bd4ibu3pugyrn6gaxw2t6ebaskr7oj44l6bvzky

DataCap allocation requested

512TiB

Id

6633d351-db72-4d3b-b98f-ca2c1507e6c4

datacap-bot[bot] commented 3 months ago

Application is ready to sign

datacap-bot[bot] commented 2 months ago

DataCap and CID Checker Report Summary[^1]

Storage Provider Distribution

✔️ Storage provider distribution looks healthy.

Deal Data Replication

✔️ Data replication looks healthy.

Deal Data Shared with other Clients[^3]

✔️ No CID sharing has been observed.

[^1]: To manually trigger this report, add a comment with text checker:manualTrigger

[^2]: Deals from those addresses are combined into this report as they are specified with checker:manualTrigger

[^3]: To manually trigger this report with deals from other related addresses, add a comment with text checker:manualTrigger <other_address_1> <other_address_2> ...

Full report

Click [here]() to view the CID Checker report.