Skip to content

Loading models is not easy with from_pretrained #188

@NamburiSrinath

Description

@NamburiSrinath

Hi team,

First off, thanks for the excellent repo and great work.

I was able to slice a model but when I am trying to load it using AutoModelForCausalLM.from_pretrained(), I am getting the following error

OSError: Error no file named pytorch_model.bin, model.safetensors, tf_model.h5, model.ckpt.index or flax_model.msgpack found in directory /xxx/TransformerCompression/experiments/llama_7b_0.25.

I've checked all the issues mentioned but am unable to find a proper fix. Please guide as it will be super helpful :)

Something like the below would be helpful as the other codebases just load the model using AutoModelForCausalLM.from_pretrained() and it's not feasible to modify all the other codebases :(

Image

Looking forward

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions