Lucain Pouget PRO
AI & ML interests
Articles
Organizations
Wauplin's activity
We've just released ๐๐๐๐๐๐๐๐๐๐๐_๐๐๐ v0.25.0 and it's packed with powerful new features and improvements!
โจ ๐ง๐ผ๐ฝ ๐๐ถ๐ด๐ต๐น๐ถ๐ด๐ต๐๐:
โข ๐ ๐จ๐ฝ๐น๐ผ๐ฎ๐ฑ ๐น๐ฎ๐ฟ๐ด๐ฒ ๐ณ๐ผ๐น๐ฑ๐ฒ๐ฟ๐ with ease using
huggingface-cli upload-large-folder
. Designed for your massive models and datasets. Much recommended if you struggle to upload your Llama 70B fine-tuned model ๐คกโข ๐ ๐ฆ๐ฒ๐ฎ๐ฟ๐ฐ๐ต ๐๐ฃ๐: new search filters (gated status, inference status) and fetch trending score.
โข โก๐๐ป๐ณ๐ฒ๐ฟ๐ฒ๐ป๐ฐ๐ฒ๐๐น๐ถ๐ฒ๐ป๐: major improvements simplifying chat completions and handling async tasks better.
Weโve also introduced tons of bug fixes and quality-of-life improvements - thanks to the awesome contributions from our community! ๐ช
๐ก Check out the release notes: Wauplin/huggingface_hub#8
Want to try it out? Install the release with:
pip install huggingface_hub==0.25.0
Thanks for the ping @clem !
This documentation is more recent regarding HfApi
(the Python client). You have methods like model_info
and list_models
to get details about models (and similarly with datasets and Spaces). In addition to the package reference, we also have a small guide on how to use it.
Otherwise, if you are interested in the HTTP endpoint to build your requests yourself, here is the API reference.
Depends what you want to do. We have full documentation here: https://huggingface.co/docs/huggingface_hub/index. You can find many guides showing you how to use the library.
Are you referring to Agents in transformers
? If yes, here is the docs about it: https://huggingface.co/docs/transformers/agents. Regarding tools, TGI supports them and the InferenceClient from huggingface_hub as well, meaning you can pass tools to chat_completion
(see "Example using tools:" section in https://huggingface.co/docs/huggingface_hub/v0.24.0/en/package_reference/inference_client#huggingface_hub.InferenceClient.chat_completion). These tools parameters were already available on huggingface_hub 0.23.x.
Hope this answers your question :)
Exciting updates include:
โก InferenceClient is now a drop-in replacement for OpenAI's chat completion!
โจ Support for response_format, adapter_id , truncate, and more in InferenceClient
๐พ Serialization module with a save_torch_model helper that handles shared layers, sharding, naming convention, and safe serialization. Basically a condensed version of logic scattered across safetensors, transformers , accelerate
๐ Optimized HfFileSystem to avoid getting rate limited when browsing HuggingFaceFW/fineweb
๐จ HfApi & CLI improvements: prevent empty commits, create repo inside resource group, webhooks API, more options in the Search API, etc.
Check out the full release notes for more details:
Wauplin/huggingface_hub#7
๐
Mostly that it's better integrated with HF services. If you pass a model_id
you can use the serverless Inference API without setting an base_url
. No need either to pass an api_key
if you are already logged in (with $HF_TOKEN
environment variable or huggingface-cli login
). If you are an Inference Endpoint user (i.e. deploying a model using https://ui.endpoints.huggingface.co/), you get a seamless integration to make requests to it with URL already configured. Finally, you are assured that the client will stay up to date with latest updates in TGI/Inference API/Inference Endpoints.
Why use the InferenceClient?
๐ Seamless transition: keep your existing code structure while leveraging LLMs hosted on the Hugging Face Hub.
๐ค Direct integration: easily launch a model to run inference using our Inference Endpoint service.
๐ Stay Updated: always be in sync with the latest Text-Generation-Inference (TGI) updates.
More details in https://huggingface.co/docs/huggingface_hub/main/en/guides/inference#openai-compatibility
Exciting updates include:
๐ Seamless download to local dir!
๐ก Grammar and Tools in InferenceClient!
๐ Documentation full translated to Korean!
๐ฅ User API: get likes, upvotes, nb of repos, etc.!
๐งฉ Better model cards and encoding for ModelHubMixin!
Check out the full release notes for more details:
Wauplin/huggingface_hub#6
๐
Thanks for your help and dedication on that class @not-lain !
huggingface_hub
Python library!Exciting updates include:
โจ Chat-completion API in the InferenceClient!
๐ค Official inference types in InferenceClient!
๐งฉ Better config and tags in
ModelHubMixin
!๐ Generate model cards for your
ModelHubMixin
integrations! ๐๏ธ x3 download speed in
HfFileSystem
!!Check out the full release notes for more details: Wauplin/huggingface_hub#5 ๐
Thanks for the kind words
@tonyassi
โค๏ธ
We're always trying to improve our tooling so feel free to reach out on https://github.com/huggingface/huggingface_hub if you have any feedback ๐ค
huggingface_hub
Python library!Exciting updates include:
๐๏ธ Dataclasses everywhere for improved developer experience!
๐พ HfFileSystem optimizations!
๐งฉ
PyTorchHubMixin
now supports configs and safetensors!โจ
audio-to-audio
supported in the InferenceClient!๐ Translated docs in Simplified Chinese and French!
๐ Breaking changes: simplified API for listing models and datasets!
Check out the full release notes for more details: Wauplin/huggingface_hub#4 ๐ค๐ป