Skip to content

[converter] Quantization option for converters.convert_jax #7260

@josephrocca

Description

@josephrocca

System information

  • TensorFlow.js version (you are using): latest
  • Are you willing to contribute it: Unfortunately, it is beyond me.

The title is pretty self explanatory, but one thing: I really hope if this is implemented, it will support converting and quantizing models above 2GB. Currently if I try to convert a 3GB model with converters.convert_jax, I get this error:

ValueError: Message tensorflow.GraphDef exceeds maximum protobuf size of 2GB: 3450353084

During handling of the above exception, another exception occurred:

AssertionError                            Traceback (most recent call last)
[/usr/local/lib/python3.8/dist-packages/tensorflow/python/framework/graph_util_impl.py](https://localhost:8080/#) in _assert_nodes_are_present(name_to_node, nodes)
    202   """Assert that nodes are present in the graph."""
    203   for d in nodes:
--> 204     assert d in name_to_node, "%s is not in graph" % d
    205 
    206

Of course, 3GB is a bit too big for the web, but 700mb is fine. About the same size as the combined text+image CLIP models used in this project which runs fine in the browser. So although 2GB seems like "more than enough" for any reasonable use case, please keep in mind that it actually limits models to ~500mb, in the sense that I've just described.

Metadata

Metadata

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions