AI & ML interests

LLMs, NLP, Alignment, DPO, RLHF, data labeling, text-classification, text-generation, token-classification

argilla 's collections 10

Synthetic Data Generator
A collection of tools and datasets related to no-code the Synthetic Data Generation.
Open Image Generation Models
A collection of models that are open source equivalents of flux-schnell and flux-dev.
Notus 7B v1
Notus 7B v1 models (DPO fine-tune of Zephyr SFT) and datasets used. More information at https://github.com/argilla-io/notus
DIBT Prompt collective SPIN
This collection contains resources related to the replication of SPIN with the dibt prompt collective dataset
Preference Datasets for KTO
This collection contains a list of curated preference datasets for KTO fine-tuning for intent alignment of LLMs through signals.
Datasets built with βš—οΈ distilabel
This collection contains some datasets generated and/or labelled using https://github.com/argilla-io/distilabel
Argilla v2.0 compatible datasets
Ready for rg.Dataset.from_hub(). Each dataset contains a my_dataset_name/tree/main/creation_script.py to see the fullconfig and creation pipeline.
Notux 8x7B v1
Notux 8x7B v1 model (DPO fine-tune of Mixtral 8x7B Instruct v0.1) and datasets used. More information at https://github.com/argilla-io/notus
Preference Datasets for DPO
This collection contains a list of curated preference datasets for DPO fine-tuning for intent alignment of LLMs
Synthetic Data Generator
A collection of tools and datasets related to no-code the Synthetic Data Generation.
Datasets built with βš—οΈ distilabel
This collection contains some datasets generated and/or labelled using https://github.com/argilla-io/distilabel
Open Image Generation Models
A collection of models that are open source equivalents of flux-schnell and flux-dev.
Argilla v2.0 compatible datasets
Ready for rg.Dataset.from_hub(). Each dataset contains a my_dataset_name/tree/main/creation_script.py to see the fullconfig and creation pipeline.
Notus 7B v1
Notus 7B v1 models (DPO fine-tune of Zephyr SFT) and datasets used. More information at https://github.com/argilla-io/notus
Notux 8x7B v1
Notux 8x7B v1 model (DPO fine-tune of Mixtral 8x7B Instruct v0.1) and datasets used. More information at https://github.com/argilla-io/notus
DIBT Prompt collective SPIN
This collection contains resources related to the replication of SPIN with the dibt prompt collective dataset
Preference Datasets for DPO
This collection contains a list of curated preference datasets for DPO fine-tuning for intent alignment of LLMs
Preference Datasets for KTO
This collection contains a list of curated preference datasets for KTO fine-tuning for intent alignment of LLMs through signals.