You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In general model weights on huggingface are a bit of a mess because of different implementations in ML libraries. For example, tinygrad implementation of models name things slightly differently to MLX implementation, which names things slightly different to torch implementation
This means we need to have some code that "converts" these names / structure to the tinygrad one
Right now there's some code that already does this to convert from the huggingface torch implementation to tinygrad:
The text was updated successfully, but these errors were encountered:
AlexCheema
changed the title
Support MLX community models in tinygrad inference engine
[BOUNTY - $100] Support MLX community models in tinygrad inference engine
Sep 5, 2024
AlexCheema
changed the title
[BOUNTY - $100] Support MLX community models in tinygrad inference engine
[BOUNTY - $200] Support MLX community models in tinygrad inference engine
Sep 5, 2024
Does this bounty also requires porting mlx modelling code to tinygrad? Since according to mlx-examples library, different models on mlx-community requires different modelling code.
exo currently only has llama. llama tinygrad modelling code is incompatible (different) with weights from qwen, etc.
exo/exo/inference/tinygrad/models/llama.py
Lines 220 to 249 in 41f0a22
The text was updated successfully, but these errors were encountered: