Open martapiekarska opened 2 months ago
Application is waiting for allocator review
@solopine The Site you provided seems to be working only on the home page. Is this the actual page? All of SPs are in the same region. You asked for 1 PiB data, while you claim you will have 3 replicas with a single dataset of 18 GiB. What do you need the extra Data Cap for? The data sample you provided is not suitable for review. Why are you applying as another organisation (as far as I know, this is your previous application)?
Thank you for review my application.
KYC has been requested. Please complete KYC at https://kyc.allocator.tech/?owner=fidlabs&repo=Enterprise-Data-Pathway&client=f1l54fu3pign7z5b3gopbequyjhn4d2ogac4pg56a&issue=69
@solopine thank you for the explanation.
The next steps required are: -Github User KYC Check via Gitcoin passport -KYB Form (LINK)
Once completed, you will be eligible for 50TiBs of DataCap. We will review your use of this amount for the next allocation and consider enlarging it if your action is good.
KYC completed for client address f1l54fu3pign7z5b3gopbequyjhn4d2ogac4pg56a
with Optimism address 0xc156dfde48c5eBD247223AfFF7cbe6887e098437
and passport score 23
.
@martplo, thank you. I have completed KYC, and submitted on https://form.jotform.com/242733516818057,and in the submission I replaced 2 SPs in Hong Kong with another 2 SPs in San Francisco.
KYB is not workable on https://efilplus.synaps.me/networks/signup, and I am wait for the bug-fix.
We have completed KYB by video call interview, and signed a contract with a SP company to store our data.
Everything has been confirmed. KYB/KYC processes were passed. Contract with the SP confirmed.
1st tranche (5%) will be issued.
Total DataCap requested
1 PiB
Expected weekly DataCap usage rate
DataCap Amount - First Tranche
50TiB
Client address
f1l54fu3pign7z5b3gopbequyjhn4d2ogac4pg56a
f1l54fu3pign7z5b3gopbequyjhn4d2ogac4pg56a
50TiB
ec53d446-6745-41fe-beae-73385f8f044a
Application is ready to sign
Your Datacap Allocation Request has been approved by the Notary
bafy2bzacebksqwyfd57d2cdvktp4onbj5xpwcbuv2domjtkklillcyhcx2fdq
Address
f1l54fu3pign7z5b3gopbequyjhn4d2ogac4pg56a
Datacap Allocated
50TiB
Signer Address
f1msap4wvgzzv4xlzeq6kycmgx55ferfloxnt2rcy
Id
ec53d446-6745-41fe-beae-73385f8f044a
You can check the status of the message here: https://filfox.info/en/message/bafy2bzacebksqwyfd57d2cdvktp4onbj5xpwcbuv2domjtkklillcyhcx2fdq
Application is Granted
(OLD vs NEW)
Weekly Allocation: 200TiB vs State: ChangesRequested vs Granted
checker:manualTrigger
No application info found for this issue on https://datacapstats.io/clients.
[^1]: To manually trigger this report, add a comment with text checker:manualTrigger
checker:manualTrigger
No active deals found for this client.
[^1]: To manually trigger this report, add a comment with text checker:manualTrigger
Sorry to late response, we are struggling to solve some business matters with our SP company, and also we're experiencing some technical issues in the process of packaging and transferring data to our SP partner. It's the first time for us to do the business, a little more complicated than expected. So can you give me more time (at most 2 weeks) to complete the first 50tb data storage? Thank you!
Client used 75% of the allocated DataCap. Consider allocating next tranche.
checker:manualTrigger
⚠️ 1 storage providers sealed too much duplicate data - f03143705: 23.02%
✔️ Data replication looks healthy.
✔️ No CID sharing has been observed.
[^1]: To manually trigger this report, add a comment with text checker:manualTrigger
[^2]: Deals from those addresses are combined into this report as they are specified with checker:manualTrigger
[^3]: To manually trigger this report with deals from other related addresses, add a comment with text checker:manualTrigger <other_address_1> <other_address_2> ...
Click here to view the CID Checker report.
checker:manualTrigger
⚠️ 2 storage providers sealed too much duplicate data - f03143698: 20.13%, f03143705: 23.31%
⚠️ 100.00% of Storage Providers have retrieval success rate equal to zero.
⚠️ 100.00% of Storage Providers have retrieval success rate less than 75%.
⚠️ The average retrieval success rate is 0.00%
✔️ Data replication looks healthy.
✔️ No CID sharing has been observed.
[^1]: To manually trigger this report, add a comment with text checker:manualTrigger
[^2]: Deals from those addresses are combined into this report as they are specified with checker:manualTrigger
[^3]: To manually trigger this report with deals from other related addresses, add a comment with text checker:manualTrigger <other_address_1> <other_address_2> ...
Click here to view the CID Checker report.
For duplicate issue, I don't know which granularity level the duplication exists. I checked our data: every piece is different, every block in the piece is different. However, in data row level, every row of one data group contains the same market security symbol, because these data is market tick data which must contain the same symbol data for one group data.
For spark retreive rate, I think next week the rate will be high because we just start sealing from this week.
Due to reporting issue, we'll grant another 50TiB till the reporting issue is not figured.
Application is in Refill
Your Datacap Allocation Request has been approved by the Notary
bafy2bzacea223xxn6v4w7feavzhnrxihbopu3npf54csvpvmcpw3qspdnyzxu
Address
f1l54fu3pign7z5b3gopbequyjhn4d2ogac4pg56a
Datacap Allocated
50TiB
Signer Address
f1msap4wvgzzv4xlzeq6kycmgx55ferfloxnt2rcy
Id
c8f8765b-7928-4e70-91c4-05181c89bd40
You can check the status here https://filfox.info/en/message/bafy2bzacea223xxn6v4w7feavzhnrxihbopu3npf54csvpvmcpw3qspdnyzxu
Application is Granted
Client used 75% of the allocated DataCap. Consider allocating next tranche.
checker:manualTrigger
✔️ Storage provider distribution looks healthy.
⚠️ 100.00% of Storage Providers have retrieval success rate equal to zero.
⚠️ 100.00% of Storage Providers have retrieval success rate less than 75%.
⚠️ The average retrieval success rate is 0.00%
✔️ Data replication looks healthy.
✔️ No CID sharing has been observed.
[^1]: To manually trigger this report, add a comment with text checker:manualTrigger
[^2]: Deals from those addresses are combined into this report as they are specified with checker:manualTrigger
[^3]: To manually trigger this report with deals from other related addresses, add a comment with text checker:manualTrigger <other_address_1> <other_address_2> ...
Click here to view the CID Checker report.
We are working together with our SPs to see why retrieval rate is zero. All SP nodes can be found on https://api.filspark.com/, and all data block can be dowloaded by manual. We are studying why spark cannot find our data.
@solopine, thank you. The reports are fixed now, so the only thing left is to confirm the retrieval rate. I don't see any reason why we shouldn't continue with the agreed schedule, so let me know when the retrieval is sorted, and I'll grant you 156TiB, finalising the first tranche of 256TiB.
Thank you for your patience, and give me a chance to investigate and fix the retrieval rate issue. I will update here when we fix it. I think it should be soon.
@solopine Given our good cooperation so far and your high commitment to quality and good behaviour, we have decided to give you another allocation, trusting that you will work to improve your retrieval results.
However, I made a mistake in calculations, and the first tranche should be 50 TiB (5%) and the second 150 TiB (15%), so I will award another 100 TiB, fulfilling the conditions of the second tranche.
Application is in Refill
Your Datacap Allocation Request has been approved by the Notary
bafy2bzacedcfdwdzgfrg3kp7f5su6fos57ijrcvaeoiqn6pba47ze6k4i5bii
Address
f1l54fu3pign7z5b3gopbequyjhn4d2ogac4pg56a
Datacap Allocated
100TiB
Signer Address
f1msap4wvgzzv4xlzeq6kycmgx55ferfloxnt2rcy
Id
ad0fa2bf-12b8-4e6b-9bf5-b0a9cea120db
You can check the status here https://filfox.info/en/message/bafy2bzacedcfdwdzgfrg3kp7f5su6fos57ijrcvaeoiqn6pba47ze6k4i5bii
Application is Granted
@martplo, Thank you for your trust. For spark retreival rate, I asked in Slack, the spark teammember said "Every 20-minute round, there is approx. ~1.6% chance that one of your deals will be included in the ~1000 deals we choose for testing". I need more time to get higher spark rate. And also we found some boost configuration issue. I think we will get higher spark rate soon.
Client used 75% of the allocated DataCap. Consider allocating next tranche.
I checked spark retrieve rate on https://grafana.filstation.app/dashboards, my 4 SPs rate is growing up from zero to about 60%. I think it will get more high as time going.
checker:manualTrigger
✔️ Storage provider distribution looks healthy.
⚠️ 100.00% of Storage Providers have retrieval success rate less than 75%.
✔️ Data replication looks healthy.
✔️ No CID sharing has been observed.
[^1]: To manually trigger this report, add a comment with text checker:manualTrigger
[^2]: Deals from those addresses are combined into this report as they are specified with checker:manualTrigger
[^3]: To manually trigger this report with deals from other related addresses, add a comment with text checker:manualTrigger <other_address_1> <other_address_2> ...
Click here to view the CID Checker report.
@solopine, I’m glad to see the improvement in retrieval performance. While there’s still room for further progress, the changes since the last report are very positive. Thank you for your efforts!
Based on this progress, I’ll grant the 3rd tranche of 300 TiB.
Application is in Refill
Version
2024-09-13T16:45:22.690Z
DataCap Applicant
@solopine
Data Owner Name
Beijing Tangxiang Technology Co., Ltd.
Data Owner Country/Region
IT & Technology Services
Website
http://www.tangshine.com
Social Media Handle
@Song (solopine@gmail.com)
Social Media Type
Slack
What is your role related to the dataset
Dataset Owner
Total amount of DataCap being requested
1 PiB
Expected size of single dataset (one copy)
18 GiB
Number of replicas to store
3
Weekly allocation of DataCap requested
200TiB
On-chain address for first allocation
f1l54fu3pign7z5b3gopbequyjhn4d2ogac4pg56a
Data Type of Application
Private Commercial/Enterprise
Identifier
Share a brief history of your project and organization
My company was established in Beijing, China in 2012. It is a certified high-tech enterprise in China. The company mainly engages in financial related software development and data services. The company's current project focuses on the development of trading platform software, transaction backtesting, transaction data test analysis, and AI trading strategy training.
Is this project associated with other projects/ecosystem stakeholders?
No
If answered yes, what are the other projects/ecosystem stakeholders
history trading data, trading strategy test and training data
Where was the data currently stored in this dataset sourced from
My Own Storage Infra
If you answered "Other" in the previous question, enter the details here
If you are a data preparer. What is your location (Country/Region)
China
If you are a data preparer, how will the data be prepared? Please include tooling used and technical details?
The trading data will be exported as csv file by company internal tool, and files will be classified by date in folder. Then the dataset file directory will be packaged in car file by our internal tool built on go-car.
If you are not preparing the data, who will prepare the data? (Provide name and business)
Has this dataset been stored on the Filecoin network before? If so, please explain and make the case why you would like to store this dataset again to the network. Provide details on preparation and/or SP distribution.
No, it hasn't.
Please share a sample of the data
"data": { "fields": [ "ts_code", "symbol", "name", "area", "industry", "fullname", "enname", "cnspell", "market", "exchange", "curr_type", "list_status", "list_date", "delist_date", "is_hs", "act_name", "act_ent_type" ], "items": [ [ "600001.SH", "600001", "邯郸钢铁(退)", null, null, "邯郸钢铁股份有限公司", "Handan Iron & Steel Co.,Ltd.", "hdgt", "主板", "SSE", "CNY", "D", "19980122", "20091229", "N", null, null ], [ "600002.SH", "600002", "齐鲁石化(退)", null, null, "中国石化齐鲁股份有限公司", "Sinopec Qilu Company Ltd.", "qlsh", "主板", "SSE", "CNY", "D", "19980408", "20060424", "N", null, null ], [ "600003.SH", "600003", "ST东北高(退)", null, null, "东北高速公路股份有限公司", "Northeast Expressway Company Ltd.", "stdbg", "主板", "SSE", "CNY", "D", "19990810", "20100226", "N", "黑龙江省交通运输厅", "地方国企" ], [ "600005.SH", "600005", "武钢股份(退)", null, null, "武汉钢铁股份有限公司", "Wuhan Iron and Steel Company Limited", "wggf", "主板", "SSE", "CNY", "D", "19990803", "20170214", "N", null, null ],[ "T00018.SH", "T00018", "上港集箱(退)", null, null, "上海港集装箱股份有限公司", "Shanghai Port Container Co., Ltd.", "sgjx", null, "SSE", "CNY", "D", "20000719", "20061020", "N", null, null ] ], "has_more": false }
Confirm that this is a public dataset that can be retrieved by anyone on the Network
Confirm
If you chose not to confirm, what was the reason
What is the expected retrieval frequency for this data
Weekly
For how long do you plan to keep this dataset stored on Filecoin
More than 3 years
In which geographies do you plan on making storage deals
Greater China, Asia other than Greater China
How will you be distributing your data to storage providers
Shipping hard drives
How did you find your storage providers
Partners
If you answered "Others" in the previous question, what is the tool or platform you used
Please list the provider IDs and location of the storage providers you will be working with.
f03143698 Hongkong, f03143705 Hongkong, f0870354 Hongkong, f01989372 Hongkong,
How do you plan to make deals to your storage providers
Boost client
If you answered "Others/custom tool" in the previous question, enter the details here
Can you confirm that you will follow the Fil+ guideline
Yes