Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SUBMISSION] MODULE 1 #112

Open
wants to merge 1 commit into
base: december-2024
Choose a base branch
from

Conversation

Knight7561
Copy link

@Knight7561 Knight7561 commented Dec 16, 2024

December 2024 Student Submission

Module Completed

  • Module 1: Instruction Tuning
  • Module 2: Preference Alignment
  • Module 3: Parameter-efficient Fine-tuning
  • Module 4: Evaluation
  • Module 5: Vision-language Models
  • Module 6: Synthetic Datasets
  • Module 7: Inference
  • Module 8: Deployment

Changes Made

Describe what you've done in this PR:

  1. What concepts did you learn?
  • Tokenisation, Reading datasets from HF, fine tuning smol models and how tokinzer module is contained
  1. What changes or additions did you make?
  • Added comments for better understanding, will add additional examples in the future work
  1. Any challenges you faced?
  • Nothing for this module

Notebooks Added/Modified

List any notebooks you've added or modified:

  • Added new example in module_name/student_examples/my_example.ipynb
  • Modified existing notebook with additional examples
  • Added documentation or comments

Checklist

  • I have read the module materials
  • My code runs without errors
  • I have pushed models and datasets to the huggingface hub
  • My PR is based on the december_2024 branch

Questions or Discussion Points

Add any questions you have or points you'd like to discuss:
1.
2.

Additional Notes

Any other information that might be helpful for reviewers:

@Knight7561 Knight7561 marked this pull request as ready for review December 16, 2024 17:06
@burtenshaw
Copy link
Collaborator

Structurally this looks great.

@ShankarChavan Could you peer-review the notebooks please?

@Knight7561
Copy link
Author

@burtenshaw @ShankarChavan Is my PR good ?

Copy link

@ShankarChavan ShankarChavan left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

In your SFT_finetuning_example the model has not shown any improvement on loss function. Also, the response from model for the given test-questions were giving repeated response as model has not learned yet. Please try to fix these issues from your end.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants