Skip to content
This repository was archived by the owner on Nov 19, 2025. It is now read-only.

Data layer rework#72

Draft
trias702 wants to merge 5 commits intomainfrom
degert/data-rework
Draft

Data layer rework#72
trias702 wants to merge 5 commits intomainfrom
degert/data-rework

Conversation

@trias702
Copy link
Copy Markdown
Collaborator

@trias702 trias702 commented Jan 5, 2024

What does this PR do ?

Rewrites the builder and dataset classes/functions to remove a lot of the superfluous holdover elements. Also unifies the RM and DPO jsonl formats, and switches the RM model to use GBS batching similar to DPO.

Changelog

  • Please update the CHANGELOG.md under next version with high level changes in this PR.

Usage

  • You can potentially add a usage example below
# Add a code snippet demonstrating how to use this 

Before your PR is "Ready for review"

Pre checks:

Checklist when contributing a new algorithm

  • Does the trainer resume and restore model state all states?
  • Does the trainer support all parallelism techniques(PP, TP, DP)?
  • Does the trainer support max_steps=-1 and validation?
  • Does the trainer only call APIs defined in alignable_interface.py?
  • Does the trainer have proper logging?

Additional Information

  • Related to # (issue)

Signed-off-by: Daniel Egert <degert@nvidia.com>
Signed-off-by: Daniel Egert <degert@nvidia.com>
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant