hack_ml | 1 year ago | on: Mistral OCR
hack_ml's comments
hack_ml | 1 year ago | on: Zamba2-7B
hack_ml | 1 year ago | on: cuDF – GPU DataFrame Library
HoloViews hvPlot Datashader Plotly Bokeh Seaborn Panel PyDeck cuxfilter node RAPIDS
hack_ml | 1 year ago | on: cuDF – GPU DataFrame Library
hack_ml | 1 year ago | on: GPT-4o
hack_ml | 2 years ago | on: Nemotron-4 15B large multilingual language model trained on 8T tokens
introduce Nemotron-4 15B, a 15-billion-parameter large multilingual language model trained on 8 trillion text tokens. Nemotron-4 15B demonstrates strong performance when assessed on English, multilingual, and coding tasks: it outperforms all existing similarly-sized open models on 4 out of 7 downstream evaluation areas and achieves competitive performance to the leading open models in the remaining ones. Specifically, Nemotron-4 15B exhibits the best multilingual capabilities of all similarly-sized models, even outperforming models over four times larger and those explicitly specialized for multilingual tasks.
hack_ml | 3 years ago | on: BERTopic: The Future of Topic Modeling
Checkout the docs at: https://maartengr.github.io/BERTopic/faq.html#can-i-use-the-...
All you need to do is below
from bertopic import BERTopic
from cuml.cluster import HDBSCAN
from cuml.manifold import UMAP
# Create instances of GPU-accelerated UMAP and HDBSCAN
umap_model = UMAP(n_components=5, n_neighbors=15, min_dist=0.0)
hdbscan_model = HDBSCAN(min_samples=10, gen_min_span_tree=True)
# Pass the above models to be used in BERTopic
topic_model = BERTopic(umap_model=umap_model, hdbscan_model=hdbscan_model)
topics, probs = topic_model.fit_transform(docs)hack_ml | 4 years ago | on: Dask – A flexible library for parallel computing in Python
See dask-sql: https://dask-sql.readthedocs.io/en/latest/pages/api.html
hack_ml | 4 years ago | on: Dask – A flexible library for parallel computing in Python
hack_ml | 4 years ago | on: Intel Extension for Scikit-Learn