top | item 43798917 (no title) aazo11 | 10 months ago This is a huge unlock for on-device inference. The download time of larger models makes local inference unusable for non-technical users. discuss order hn newest No comments yet.
No comments yet.