You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This release includes general improvements to the library and new metrics within the NLP domain.
Natural language processing is arguably one of the most exciting areas of machine learning, with models such as BERT, ROBERTA, GPT-3 etc., really pushing what automated text translation, recognition, and generation systems are capable of.
With the introduction of these models, many metrics have been proposed that measure how well these models perform. TorchMetrics v0.5 includes 4 such metrics: BERT score, BLEU, ROUGE and WER.
num_thresholds argument in BinnedPrecisionRecallCurve
Fixed
Fixed bug where classification metrics with average='macro' would lead to wrong result if a class was missing (Fix metrics in macro average #303)
Fixed weighted, multi-class AUROC computation to allow for 0 observations of some class, as contribution to final AUROC is 0 (Weighted AUROC to omit empty classes #376)
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
[0.5.0] - 2021-08-09
This release includes general improvements to the library and new metrics within the NLP domain.
Natural language processing is arguably one of the most exciting areas of machine learning, with models such as BERT, ROBERTA, GPT-3 etc., really pushing what automated text translation, recognition, and generation systems are capable of.
With the introduction of these models, many metrics have been proposed that measure how well these models perform. TorchMetrics v0.5 includes 4 such metrics: BERT score, BLEU, ROUGE and WER.
Detail changes
Added
MetricTracker
wrapper metric for keeping track of the same metric over multiple epochs (metric tracker #238)nDCG
metric for target with values larger than 1 (Allow target nDCG metric to be integer larger than 1 #349)nDCG
metric (fix nDCG can not be called with negative relevance targets #378)None
as reduction option inCosineSimilarity
metric (Add None as reduction option in CosineSimilarity #400)AveragePrecision
(multilabel for AveragePrecision #386)Changed
psnr
andssim
fromfunctional.regression.*
tofunctional.image.*
(move functional psnr & ssim to image #382)image_gradient
fromfunctional.image_gradients
tofunctional.image.gradients
(Move image gradient #381)R2Score
fromregression.r2score
toregression.r2
(cleaning & prune re-definine #371)torch.argmax
instead oftorch.topk
whenk=1
for better performance (Use argmax when topk=1 #419)Deprecated
r2score
>>r2_score
andkldivergence
>>kl_divergence
infunctional
(cleaning & prune re-definine #371)bleu_score
fromfunctional.nlp
tofunctional.text.bleu
(Added Blue Score the respective folders #360)Removed
threshold
has to be in (0,1) range to support logit input (Allow threshold to be outside (0,1) domain #351, Remove remaining threshold checks #401)preds
could not be bigger thannum_classes
to support logit input (Remove check that preds value need to be smaller than num_classes #357)regression.psnr
andregression.ssim
(move functional psnr & ssim to image #382):functional.mean_relative_error
num_thresholds
argument inBinnedPrecisionRecallCurve
Fixed
average='macro'
would lead to wrong result if a class was missing (Fix metrics in macro average #303)weighted
,multi-class
AUROC computation to allow for 0 observations of some class, as contribution to final AUROC is 0 (Weighted AUROC to omit empty classes #376)_forward_cache
and_computed
attributes are also moved to the correct device if metric is moved (Move forward cache and computed to device #413)IoU
metric when usingignore_index
argument (fix ignore_index in the computation of IoU #328)Contributors
@BeyondTheProof, @Borda, @CSautier, @discort, @edwardclem, @gagan3012, @hugoperrin, @karthikrangasai, @paul-grundmann, @quancs, @rajs96, @SkafteNicki, @vatch123
If we forgot someone due to not matching commit email with GitHub account, let us know :]
This discussion was created from the release Text-related (NLP) metrics.
Beta Was this translation helpful? Give feedback.
All reactions