Closed alberttwong closed 4 months ago
root@spark:/opt/LakeView# cat delta.yaml
version: V1
onehouseClientConfig:
# can be obtained from the Onehouse console
projectId: c3eb3868-6979-41cd-9018-952d29a43337
apiKey: XXXXXX
apiSecret: YYYYY
userId: x2gblCN8xNSurvCsqDaGJ84zy913
file: delta.yaml
fileSystemConfiguration:
# Provide either s3Config or gcsConfig
s3Config:
region: us-west-2
accessKey: admin
accessSecret: password
metadataExtractorConfig:
jobRunMode: ONCE
pathExclusionPatterns:
parserConfig:
- lake: <lake1>
databases:
- name: people
basePaths: ["s3://warehouse/people"]
# Add additional lakes and databases as needed
If you remove version: V1
from yaml you get a different error
root@spark:/opt/LakeView# java -jar LakeView-release-v0.10.0-all.jar -p '/opt/LakeView/delta.yaml'
16:43:42.637 [main] INFO com.onehouse.Main - Starting LakeView extractor service
Exception in thread "main" java.lang.RuntimeException: Failed to load config
at com.onehouse.config.ConfigLoader.loadConfigFromConfigFile(ConfigLoader.java:31)
at com.onehouse.Main.loadConfig(Main.java:92)
at com.onehouse.Main.start(Main.java:56)
at com.onehouse.Main.main(Main.java:41)
Caused by: java.lang.NullPointerException
at com.onehouse.config.ConfigLoader.loadConfigFromJsonNode(ConfigLoader.java:44)
at com.onehouse.config.ConfigLoader.loadConfigFromConfigFile(ConfigLoader.java:29)
... 3 more
answer:
had to remove file:delta.yaml field
comments from dev:
the file option is to be used only when we dont want to pass the creds directly in the config.yaml file
environment: docker compose with openjdk 11, minio, xtable, spark 3.4, hive 2.3.10, hadoop 2.10.2