Closed smallzhongfeng closed 10 months ago
I deployed it using the latest online version, and the client is also the latest version 0.11.0
@thinkharderdev @yahoNanJing @Dandandan Have you ever encountered similar problems? Could you guys give me some advice
Similar issue like this: https://github.com/apache/arrow-ballista/issues/747
use datafusion::arrow::datatypes::DataType;
use datafusion::datasource::file_format::parquet::DEFAULT_PARQUET_EXTENSION;
use ballista::prelude::{BallistaConfig, BallistaContext, Result};
use datafusion::prelude::{CsvReadOptions, ParquetReadOptions, SessionContext};
#[tokio::main]
async fn main() -> Result<()> {
let config = BallistaConfig::builder()
.set("ballista.shuffle.partitions", "1")
.build()?;
let ctx = BallistaContext::standalone(&config, 2).await?;
let options = ParquetReadOptions {
file_extension: DEFAULT_PARQUET_EXTENSION,
table_partition_cols: vec![("date".to_string(), DataType::Utf8)],
parquet_pruning: Some(false),
skip_metadata: Some(true),
};
let path= format!("tmp");
let arc = ctx.read_parquet(&path, options).await?;
println!("{}", arc.schema());
arc.clone().select_columns(&["String", "date"]).unwrap();
arc.clone().show().await?;
Ok(())
}
This case also fail, so is it currently not supported to create a partition table?
Hi @smallzhongfeng, I'll take a look at this issue in this week.
Thank you for your reply. @yahoNanJing At present, my guess is that the partition field is treated as an ordinary field, resulting in an error when the schema is matched.
Any update ?
It looks the partitions are ignored, and the files inside are not loaded. Is there any update on how to deal that?
Any update?
Describe the bug After the partition table is created, it cannot be read normally
To Reproduce
run in ballista-cli
I deployed it in standalone mode.
Expected behavior A clear and concise description of what you expected to happen.
Additional context Add any other context about the problem here.