package tokenizer import ( "strings" "testing" ) func TestLoadFromBytesRejectsWordPiece(t *testing.T) { data := []byte(`{ "model": { "type": "WordPiece", "vocab": {"[UNK]": 0, "hello": 1} }, "added_tokens": [] }`) _, err := LoadFromBytes(data) if err == nil { t.Fatal("expected WordPiece load to fail") } if !strings.Contains(err.Error(), "unsupported tokenizer type: WordPiece") { t.Fatalf("unexpected error: %v", err) } }