Reverse Dependencies of sacrebleu
The following projects have a declared dependency on sacrebleu:
- adapter-transformers — A friendly fork of HuggingFace's Transformers, adding Adapters to PyTorch language models
- adapters — A Unified Library for Parameter-Efficient and Modular Transfer Learning
- adaptor — Adaptor: Objective-centric Adaptation Framework for Language Models.
- ai-metrics — A library for basic NLP metric score implementations
- ai2-catwalk — A library for evaluating language models.
- aigc-evals — aigc_evals
- alexa-teacher-models — Alexa Teacher Models
- arize — A helper library to interact with Arize AI APIs
- audiossl — no summary
- autogluon-contrib-nlp — MXNet GluonNLP Toolkit (DeepNumpy Version)
- AutoRAG — Automatically Evaluate RAG pipelines with your own data. Find optimal structure for new RAG product.
- bleuscore — A fast bleu score calculator
- classy-core — A powerful tool to train and use your classification models.
- cody-adapter-transformers — A friendly fork of HuggingFace's Transformers, adding Adapters to PyTorch language models
- commode-utils — Set of useful functions and modules for Code Modeling
- compare-mt — Holistic comparison of the output of text generation models
- Comprehensive-RAG-Evaluation-Metrics — This library provides a comprehensive suite of metrics to evaluate the performance of Retrieval-Augmented Generation (RAG) systems. RAG systems, which combine information retrieval with text generation, present unique evaluation challenges beyond those found in standard language generation tasks
- compression-distance — A compression-based edit distance metric for text comparison
- compromise-marian — Marian model but with two decoders
- crfm-helm — Benchmark for language models
- DashAI — DashAI: a graphical toolbox for training, evaluating and deploying state-of-the-art AI models.
- datalabs — Datalabs
- deepa2 — Cast NLP data as multiangular DeepA2 datasets and integrate these in training pipeline
- delta-nlp — DELTA is a deep learning based natural language and speech processing platform.
- dialogues — This package provides a unified interface to several dialogue benchmarks
- eole — Open language modeling toolkit based on PyTorch
- espnet — ESPnet: end-to-end speech processing toolkit
- eval-mm — eval-mm is a tool for evaluating Multi-Modal Large Language Models.
- evalRagPk — A library for evaluating Retrieval-Augmented Generation (RAG) systems
- evals — no summary
- evals-nightly — no summary
- evalscope — EvalScope: Lightweight LLMs Evaluation Framework
- evaluate — HuggingFace community-driven open-source library of evaluation
- examinationrag — XRAG: eXamining the Core - Benchmarking Foundational Component Modules in Advanced Retrieval-Augmented Generation
- explainaboard — Explainable Leaderboards for Natural Language Processing
- fairseq — Facebook AI Research Sequence-to-Sequence Toolkit
- fairseq2 — FAIR Sequence Modeling Toolkit
- fairseq2-patch — FAIR Sequence Modeling Toolkit
- flagai — FlagAI aims to help researchers and developers to freely train and test large-scale models for NLP/CV/VL tasks.
- flexeval — no summary
- flexrag — A RAG Framework for Information Retrieval and Generation.
- formerbox — no summary
- gem-metrics-fork — GEM Challenge metrics
- generate-sequences — no summary
- genienlp — no summary
- geniusrise-text — Text bolts for geniusrise
- geniusrise-vision — Huggingface bolts for geniusrise
- git-t5 — Open source machine learning framework for training T5 models on source code in JAX/Flax.
- guardrails-ai-unbabel-comet — High-quality Machine Translation Evaluation
- h2ogpt — no summary
- indic-eval — A package to make LLM evaluation easier
- IndicTrans2 — Indic NLP package
- janus-llm — A transcoding library using LLMs.
- joeynmt — Minimalist NMT for educational purposes
- jrvc — Libraries for RVC inference
- jury — Evaluation toolkit for neural language generation.
- kogito — A Python NLP Commonsense Knowledge Inference Toolkit
- kogitune — The Kogitune 🦊 LLM Project
- LangRAGEval — LangRAGEval is a library for evaluating responses based on faithfulness, context recall, answer relevancy, and context relevancy.
- lens-metric — A learnable evaluation metric for text simplification
- lighteval — A lightweight and configurable evaluation package
- lightyear — lightyear
- lit-nlp — Language Interpretability Tool.
- llm-blender — LLM-Blender, an innovative ensembling framework to attain consistently superior performance by leveraging the diverse strengths and weaknesses of multiple open-source large language models (LLMs). LLM-Blender cut the weaknesses through ranking and integrate the strengths through fusing generation to enhance the capability of LLMs.
- llmuses — Eval-Scope: Lightweight LLMs Evaluation Framework
- lm-eval — A framework for evaluating language models
- lm-polygraph — Uncertainty Estimation Toolkit for Transformer Language Models
- lmms-eval — A framework for evaluating large multi-modality language models
- LocalCat — Fine-tune Large Language Models locally.
- luna-nlg — Source code for the LUNA project
- mammoth-nlp — Massively Multilingual Modular Open Translation @ Helsinki
- MarkLLM — MarkLLM: An Open-Source Toolkit for LLM Watermarking
- mbr — Minimum Bayes risk decoding for Hugging Face Transformers
- mbrs — A library for minimum Bayes risk (MBR) decoding.
- mlbench-core — A public and reproducible collection of reference implementations and benchmark suite for distributed machine learning systems.
- modelscope — ModelScope: bring the notion of Model-as-a-Service to life.
- ms-opencompass — A lightweight toolkit for evaluating LLMs based on OpenCompass.
- mt-telescope — A visual platform for contrastive evaluation of machine translation systems
- mteval — Library to automate machine translation evaluation
- Multimodal-Keras-Wrapper — Wrapper for Keras with support to easy multimodal data and models loading and handling.
- mw-adapter-transformers — A friendly fork of HuggingFace's Transformers, adding Adapters to PyTorch language models
- narg2p — Non AutoRegressive Grapheme to Phoneme conversion Toolkit
- nemo-toolkit — NeMo - a toolkit for Conversational AI
- neurst — Neural Speech Translation Toolkit
- nmtlab — A simple framework for neural machine translation based on PyTorch
- nmtpytorch — Sequence-to-Sequence Framework in Pytorch
- opencompass — A comprehensive toolkit for large model evaluation
- OpenNMT-py — A python implementation of OpenNMT
- OpenNMT-tf — Neural machine translation and sequence learning using TensorFlow
- paddlespeech — Speech tools and models based on Paddlepaddle
- pangeamt-tea — TEA - Translation Engine Architect
- paraphrase-metrics — Paraphrase Metrics
- pdexplorer — A Stata emulator for Python/pandas
- pgml-extension — Simple machine learning in PostgreSQL.
- prompt2model — A library for distilling models from prompts.
- pureml-llm — no summary
- pytext-nlp — pytorch modeling framework and model zoo for text models
- rag-eval-tool — A comprehensive evaluation toolkit for RAG and LLMs.
- rag-evaluator — A library for evaluating Retrieval-Augmented Generation (RAG) systems
- RAGchain — Build advanced RAG workflows with LLM, compatible with Langchain
1
2