Skip to content

Conversation

@ggerganov
Copy link
Member

Add general.file_type meta info to indicate the quantization strategy used when creating the model

If found, we print it directly. Otherwise, we try to guess the file type based on the most used tensor type

@ggerganov ggerganov requested a review from ikawrakow August 22, 2023 11:42
@slaren
Copy link
Member

slaren commented Aug 22, 2023

Should this be set to F16 or F32 on convert.py?

@klosax
Copy link
Contributor

klosax commented Aug 22, 2023

Should this be set to F16 or F32 on convert.py?

Depending on what format is output, I suggest:
All tensors F32
All tensors F16

@ggerganov ggerganov merged commit deb7dfc into master Aug 22, 2023
@ggerganov ggerganov deleted the add-ftype branch August 22, 2023 17:06
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants