Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SUBMISSION] December 2024 - Module 1: Instruction tuning #78

Open
wants to merge 1 commit into
base: december-2024
Choose a base branch
from

Conversation

jasoriya
Copy link
Collaborator

@jasoriya jasoriya commented Dec 10, 2024

December 2024 Student Submission

Module Completed

  • Module 1: Instruction Tuning
  • Module 2: Preference Alignment
  • Module 3: Parameter-efficient Fine-tuning
  • Module 4: Evaluation
  • Module 5: Vision-language Models
  • Module 6: Synthetic Datasets
  • Module 7: Inference
  • Module 8: Deployment

Changes Made

Describe what you've done in this PR:

  1. What concepts did you learn?
  2. What changes or additions did you make?
  3. Any challenges you faced?

Notebooks Added/Modified

List any notebooks you've added or modified:

  • Added new example in module_name/students/my_example.ipynb
  • Modified existing notebook with additional examples
  • Added documentation or comments

Checklist

  • I have read the module materials
  • My code runs without errors
  • I have pushed models and datasets to the huggingface hub
  • My PR is based on the december_2024 branch

Questions or Discussion Points

Add any questions you have or points you'd like to discuss:
1.
2.

Additional Notes

Any other information that might be helpful for reviewers:
I tried to see how a challenging task like chain of thought prompting could be fine-tuned on a small LLM. As can be seen in the example output in the notebook, the fine-tuned model tries to start the reasoning chain but starts hallucinating soon after.

@jasoriya jasoriya changed the title Complete exercise with finetuning on COT dataset [SUBMISSION] December 2024 - Module 1: Instruction tuning Dec 10, 2024
@burtenshaw
Copy link
Collaborator

Any other information that might be helpful for reviewers:
I tried to see how a challenging task like chain of thought prompting could be fine-tuned on a small LLM. As can be seen in the example output in the notebook, the fine-tuned model tries to start the reasoning chain but starts hallucinating soon after.

This is really cool!

Start a thread on Discord if you want to discuss! https://discord.com/channels/879548962464493619/1313889336907010110

@burtenshaw
Copy link
Collaborator

Could you exchange a review with one of the other students's PRs? [SUBMISSION]

@jasoriya
Copy link
Collaborator Author

@burtenshaw I'd be glad to exchange a review. It seems that we don't have the permission to add ourselves as a reviewer. Would you be adding us as a reviewer?

@burtenshaw
Copy link
Collaborator

@burtenshaw I'd be glad to exchange a review. It seems that we don't have the permission to add ourselves as a reviewer. Would you be adding us as a reviewer?

Thanks. Just comment and mention me. Then I'll assign.

@bhautik-pithadiya
Copy link
Collaborator

@burtenshaw add me as a reviewer for this PR.

@burtenshaw
Copy link
Collaborator

@bhautik-pithadiya if you review, I'll merge.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants