Skip to content

fix(docs): correct LLM_TYPE for MiniCPM-V-4 from 'llama' to 'llama3'#1088

Open
playaswd wants to merge 1 commit intoOpenBMB:mainfrom
playaswd:fix/issue-1049-llm-type-doc-mismatch
Open

fix(docs): correct LLM_TYPE for MiniCPM-V-4 from 'llama' to 'llama3'#1088
playaswd wants to merge 1 commit intoOpenBMB:mainfrom
playaswd:fix/issue-1049-llm-type-doc-mismatch

Conversation

@playaswd
Copy link

Summary

  • Fix incorrect LLM_TYPE value in finetune documentation and shell scripts for MiniCPM-V-4
  • The docs stated LLM_TYPE="llama", but dataset.py only handles "llama3", "qwen", and defaults to "minicpm" — there is no "llama" handler
  • Users following the docs would silently get wrong tokenization (minicpm-style instead of llama3-style)

Changes

  • finetune/readme.md: llamallama3 for MiniCPM-V-4
  • finetune/finetune_ds.sh: Added MiniCPM-V-4 to the llama3 comment
  • finetune/finetune_lora.sh: Same as above

Fixes #1049

Test plan

  • Verified dataset.py:conversation_to_ids() only handles "llama3", "qwen", and default "minicpm"
  • Confirmed upstream readme.md still has the incorrect "llama" value
  • Docs now match the code behavior

🤖 Generated with Claude Code

…penBMB#1049)

The finetune documentation and shell scripts incorrectly stated that
MiniCPM-V-4 should use LLM_TYPE="llama", but the code in dataset.py
only handles "llama3", "qwen", and "minicpm" (default). There is no
"llama" handler, so users following the docs would silently get the
wrong tokenization (minicpm-style instead of llama3-style).

Since MiniCPM-V-4 uses a Llama-based LLM, the correct LLM_TYPE is
"llama3". This commit updates the documentation in readme.md,
finetune_ds.sh, and finetune_lora.sh accordingly.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[BUG] <title>

1 participant