issues
search
neulab
/
ExplainaBoard
Interpretable Evaluation for AI Systems
MIT License
360
stars
36
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Implement Serializable interface in BucketPerformance
#561
odashi
closed
1 year ago
0
Use ~= operator to declare dependencies
#560
tetsuok
closed
1 year ago
1
Remove obsolete docstring from get_processor_class
#559
tetsuok
closed
1 year ago
0
Update condition to apply Student's t-test
#558
tetsuok
closed
1 year ago
0
Reconsider default number of buckets
#557
neubig
opened
1 year ago
0
Add "create documentation" to add_tasks.md
#556
neubig
opened
1 year ago
1
Add text-to-sql evaluation metrics
#555
shuaichenchang
closed
1 year ago
7
add calibration analysis doc
#554
qjiang002
closed
1 year ago
0
file format doc with existing examples
#553
qjiang002
closed
1 year ago
0
Create pull_request_template.md
#552
odashi
closed
1 year ago
0
Set the time limit for integration tests
#551
tetsuok
closed
1 year ago
0
Add temporary fix to MetaEvalNLGCITest to fix flakiness
#550
tetsuok
closed
1 year ago
1
Accelerate ci avoid external stats
#549
pfliu-nlp
closed
1 year ago
0
fix bug: wrong error type
#548
pfliu-nlp
closed
1 year ago
0
test_generate_system_analysis in integration_tests.summarization_test.SummarizationTest is flaky
#547
tetsuok
opened
1 year ago
0
Wrap Markdown text at 88 columns
#546
tetsuok
closed
1 year ago
0
Teach markdownlint-cli2 to ignore explainaboard/third_party
#545
tetsuok
closed
1 year ago
3
Use same code block style in the same Markdown file
#544
tetsuok
closed
1 year ago
0
Specify languages in fenced code blocks in Markdown
#543
tetsuok
closed
1 year ago
1
Use headings instead of emphasis in Markdown
#542
tetsuok
closed
1 year ago
0
Enable checking the use of inline HTML in Markdown
#541
tetsuok
closed
1 year ago
0
Fix skipping heading levels in Markdown
#540
tetsuok
closed
1 year ago
0
Remove Performance
#539
odashi
closed
1 year ago
2
MetricResult.get_value_or_none
#538
odashi
closed
1 year ago
0
Potential issue with spearman R bootstrapping
#537
neubig
closed
1 year ago
10
Lint Markdown files
#536
tetsuok
closed
1 year ago
5
Add markdownlint-cli2 to pre-commit hook and CI
#535
tetsuok
closed
1 year ago
0
list[dict[Performance]] to dict[dict[Performance]]
#534
odashi
closed
1 year ago
0
Add unit test for meta_evaluation_wmt_da.
#533
odashi
closed
1 year ago
0
add third party code for text-to-sql evaluation
#532
shuaichenchang
closed
1 year ago
17
skip docstring check for third party code
#531
shuaichenchang
closed
1 year ago
1
Metrics with different analysis levels have the same name
#530
qjiang002
closed
1 year ago
5
Implement calibration analysis
#529
qjiang002
closed
1 year ago
3
Allowed specification of the metric #dimensions
#528
neubig
closed
1 year ago
12
Add task metaevaluation for nlg
#527
pfliu-nlp
closed
1 year ago
5
Introduce meta eval metric for general NLG tasks
#526
pfliu-nlp
closed
1 year ago
2
Support datalab datasets wmt metric
#525
pfliu-nlp
closed
1 year ago
0
Renaming meta evaluation
#524
pfliu-nlp
closed
1 year ago
0
fix-bug-nlg-metaevaluation
#523
pfliu-nlp
closed
1 year ago
0
Remove serializable from processors
#522
odashi
closed
1 year ago
1
Implement Loader factory
#521
odashi
closed
1 year ago
1
remove CustomFeature
#520
odashi
closed
1 year ago
0
Fix confidence level in scipy.stats.t.interval.
#519
odashi
closed
1 year ago
0
AnalysisLevel does not own distinctive information
#518
odashi
opened
1 year ago
4
Remove downloading spaCy's English pipelines from setup.py
#517
tetsuok
closed
1 year ago
1
Move configs from setup.py to setup.cfg
#516
tetsuok
closed
1 year ago
0
list[MetricSomething] to dict[str, MetricSomething]
#515
odashi
closed
1 year ago
2
Remove unnecessary Optional
#514
odashi
closed
1 year ago
4
Remove MetricResult.config.
#513
odashi
closed
1 year ago
2
Remove additional config argument from Metric
#512
odashi
closed
1 year ago
2
Previous
Next