Recent Releases of aac-metrics
aac-metrics - Version 0.5.5
[0.5.5] 2025-01-20
Added
- New
CLAPSimmetric based on the embeddings given by CLAP model. - New
MACEmetric based onCLAPSimandFERmetrics. - DCASE2024 challenge metric set, class and functions.
- Preprocess option in
evaluatenow accepts custom callable value. - List of bibtex sources in
data/papers.bibfile.
Changed
- Improve metric output typing for language servers with typed dicts.
batch_sizecan now beNoneto take all inputs at once into the model.
Fixed
bert_scoreoption in download script.
- Python
Published by Labbeti about 1 year ago
aac-metrics - Version 0.5.4
[0.5.4] 2024-03-04
Fixed
- Backward compatibility of
BERTScoreMrefswith torchmetrics prior to 1.0.0.
Deleted
Versionclass to usepackaging.version.Versioninstead.
- Python
Published by Labbeti almost 2 years ago
aac-metrics - Version 0.5.3
[0.5.3] 2024-01-09
Fixed
- Fix
BERTScoreMrefscomputation when all multiple references sizes are equal. - Check for empty timeout list in
SPICEmetric.
- Python
Published by Labbeti about 2 years ago
aac-metrics - Version 0.5.2
[0.5.2] 2024-01-05
Changed
aac-metricsis now compatible withtransformers>=4.31.- Rename default device value "auto" to "cudaifavailable".
- Python
Published by Labbeti about 2 years ago
aac-metrics - Version 0.5.1
[0.5.1] 2023-12-20
Added
- Check sentences inputs for all metrics.
Fixed
- Fix
BERTScoreMRefsmetric with 1 candidate and 1 reference.
- Python
Published by Labbeti about 2 years ago
aac-metrics - Version 0.5.0
[0.5.0] 2023-12-08
Added
- New
Vocabmetric to compute vocabulary size and vocabulary ratio. - New
BERTScoreMRefsmetric wrapper to compute BERTScore with multiple references.
Changed
- Rename metric
FluErrtoFER.
Fixed
METEORlocalization issue. (#9)SPIDErMaxoutput whenreturn_all_scores=False.
- Python
Published by Labbeti about 2 years ago
aac-metrics - Version 0.4.6
[0.4.6] 2023-10-10
Added
- Argument
clean_archivesforSPICEdownload.
Changed
- Check if newline character is in the sentences before ptb tokenization. (#6)
SPICEno longer requires bash script files for installation.
Fixed
- Maximal version of
transformersdependancy set to 4.31.0 to avoid error withFENSEandFluErrmetrics. SPICEcrash message and error output files.- Default value for
Evaluatemetricsargument.
Deleted
- Remove now useless
use_shelloption for download.
- Python
Published by Labbeti over 2 years ago
aac-metrics - Version 0.4.5
[0.4.5] 2023-09-12
Added
- Argument
use_shellforMETEORandSPICEmetrics anddownloadfunction to fix Windows-OS specific error.
Changed
- Rename
evaluate.pyscript toeval.py.
Fixed
- Workflow on main branch.
- Examples in README and doc with at least 2 sentences, and add a warning on all metrics that requires at least 2 candidates.
- Python
Published by Labbeti over 2 years ago
aac-metrics - Version 0.4.4
[0.4.4] 2023-08-14
Added
Evaluateclass now implements a__hash__andtolist()methods.BLEU1 to n classes and functions.- Get and set global user paths for cache, java and tmp.
Changed
- Function
get_install_infonow returnspackage_path. AACMetricnow indicate the output type when using__call__method.- Rename
AACEvaluatetoDCASE2023Evaluateand usedcase2023metric set instead ofallmetric set.
Fixed
sbert_simname in internal instantiation functions.- Path management for Windows.
- Python
Published by Labbeti over 2 years ago
aac-metrics - Version 0.4.3
[0.4.3] 2023-06-15
Changed
AACMetricis no longer a subclass oftorchmetrics.Metriceven when it is installed. It avoid dependency to this package and remove potential errors due to Metric.- Java 12 and 13 are now allowed.
Fixed
- Output name
sbert_simin FENSE and SBERTSim classes. Evaluateclass instantiation withtorchmetrics>= 0.11.evaluate.pyscript when using a verbose mode != 0.
- Python
Published by Labbeti over 2 years ago
aac-metrics - Version 0.4.2
[0.4.2] 2023-04-19
Fixed
- File
install_spice.shis now insrc/aac_metricsdirectory to fix download from a pip installation. (#3) - Java version retriever to avoid exception when java version is correct. (#2)
- Python
Published by Labbeti almost 3 years ago
aac-metrics - Version 0.4.1
[0.4.1] 2023-04-13
Deleted
- Old unused files
package_tree.rst,fluency_error.py,sbert.pyandspider_err.py.
- Python
Published by Labbeti almost 3 years ago
aac-metrics - Version 0.4.0
[0.4.0] 2023-04-13
Added
- Argument
return_probsfor fluency error metric.
Changed
- Rename
SPIDErErrtoSPIDErFLto match DCASE2023 metric name. - Rename
SBERTtoSBERTSimto avoid confusion with SBERT model name. - Rename
FluencyErrortoFluErr. - Check if Java executable version between 8 and 11. (#1)
Fixed
SPIDErFLsentences scores outputs when usingreturn_all_scores=True.- Argument
reset_stateinSPIDErFL,SBERTSim,FluErrandFENSEwhen using their functional interface. - Classes and functions factories now support SPICE and CIDEr-D metrics.
SBERTSimclass instantiation.
- Python
Published by Labbeti almost 3 years ago
aac-metrics - Version 0.3.0
[0.3.0] 2023-02-27
Added
- Parameters
timeoutandseparate_cache_dirinSPICEfunction and class. - Documentation pages with sphinx.
- Parameter
languageinMETEORfunction and class. - Options to download only
PTBTokenizer,METEOR,SPICEorFENSEindownload.py. SBERTandFluencyErrormetrics extracted fromFENSE.SPIDErErrmetric which combinesSPIDErwithFluencyError.- Parameter
reset_stateinSBERT,FluencyError,SPIDErErrandFENSEfunctions and classes.
Changed
- Fix README typo and SPIDEr-max tables.
Fixed
- Workflow badge with Github changes. (https://github.com/badges/shields/issues/8671)
- Python
Published by Labbeti almost 3 years ago
aac-metrics - Version 0.2.0
[0.2.0] 2022-12-14
Added
FENSEclass and function metric, with fluency error rate and raw output probabilities.- Unittest with
fenserepository. load_metricfunction in init to match huggingface evaluation package.
Changed
- Rename
global_scorestocorpus_scoresandlocal_scorestosents_scores. - Rename
CustomEvaluatetoEvaluateandcustom_evaluatetoevaluate. - Set default cache path to
$HOME/.cache. - Remove 'coco' prefix to file, functions and classes names to have cleaner names.
Fixed
FENSEmetric error when computing scores with less thanbatch_sizesentences.
- Python
Published by Labbeti about 3 years ago
aac-metrics - Version 0.1.2
[0.1.2] 2022-10-31
Added
- All candidates scores option
return_all_cands_scoresfor SPIDEr-max. - Functions
is_mono_sentsandis_mult_sentsto detectlist[str]sentences andlist[list[str]]multiples sentences. - Functions
flat_listandunflat_listto flat multiples sentences to sentences.
Changed
- Update default value used for
return_all_scoresin cider and rouge functions. - Update internal metric factory with functions instead of classes to avoid cyclic dependency.
Fixed
- Fix SPIDEr-max local scores output shape.
- Python
Published by Labbeti over 3 years ago
aac-metrics - Version 0.1.1
[0.1.1] 2022-09-30
Added
- Documentation for metric functions and classes.
- A second larger example for unit testing.
Changed
- Update README information, references and description.
Fixed
- SPIDEr-max computation with correct global and local outputs.
- Unit testing for computing SPICE metric from caption-evaluation-tools.
- Python
Published by Labbeti over 3 years ago
aac-metrics - Version 0.1.0
[0.1.0] 2022-09-28
Added
- BLEU, METEOR, ROUGE-l, SPICE, CIDEr and SPIDEr metrics functions and modules.
- SPIDEr-max experimental implementation.
- Installation script in download.py.
- Evaluation script in evaluate.py.
- Unittest with
caption-evaluation-toolsrepository.
- Python
Published by Labbeti over 3 years ago