Skip to content

Conversation

@Vaibhavs10
Copy link
Contributor

ty codex!

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.


You can easily integrate Hugging Face Hub with a variety of Identity Providers, such as Okta, OneLogin or Azure Active Directory (Azure AD). Hugging Face Hub can work with any OIDC-compliant or SAML Identity Provider.

As of November 2025, the official Hugging Face App is listed in the Okta Integration Network, providing a pre-validated SAML configuration that organizations can add directly from the network. OIDC, SCIM, and JIT provisioning are planned for the same app soon.
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

is there a link on Okta side we should point to?

If you don't have any papers on Hugging Face yet, you can index your first one as explained [here](#can-i-have-a-paper-page-even-if-i-have-no-modeldatasetspace). Once available, you can claim authorship.


## Recent additions
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
## Recent additions
## Tag your organisation on papers

maybe?

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I like this one better than "recent additions". But let's use "organization" for consistency with the following paragraph (and https://huggingface.co/docs/hub/organizations)

@Vaibhavs10
Copy link
Contributor Author

not sure if it's good idea to add some screenshots for these too!

Copy link
Member

@pcuenca pcuenca left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice rundown of recent features! I found a few of the items slightly out of place or with unclear segues, made a few suggestions.

Enterprise Hub organizations gain access to advanced compute options to accelerate their machine learning journey.

In addition to ZeroGPU, `hf jobs run…` executes arbitrary CPU or GPU workloads on Hugging Face. The Jobs service is billed by the second, supports simple scripts and large fleets of hardware, and is controlled via the `hf jobs run` CLI (see [the Jobs guide](https://huggingface.co/docs/huggingface_hub/en/guides/jobs)).
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I would add this as a mini-section after the ZeroGPU one, because we haven't said anything about ZeroGPU at this point in the doc.

<img class="hidden dark:block" src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/hub/gguf-tensortable-dark.png"/>
</div>

For GGUF files up to 10 GB the **GGUF Editor** label powered by Xet appears; clicking it opens an in-browser metadata editor so you can tweak properties such as the chat template without re-uploading the entire file.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
For GGUF files up to 10 GB the **GGUF Editor** label powered by Xet appears; clicking it opens an in-browser metadata editor so you can tweak properties such as the chat template without re-uploading the entire file.
GGUF files up to 10 GB can be edited using an in-browser **GGUF Editor** for metadata entries, powered by Xet. This allows you to tweak properties such as the chat template without re-uploading the entire file.

I find the first sentence a bit confusing.


The Model Hub is where the members of the Hugging Face community can host all of their model checkpoints for simple storage, discovery, and sharing. Download pre-trained models with the [`huggingface_hub` client library](https://huggingface.co/docs/huggingface_hub/index), with 🤗 [`Transformers`](https://huggingface.co/docs/transformers/index) for fine-tuning and other usages or with any of the over [15 integrated libraries](./models-libraries). You can even leverage [Inference Providers](/docs/inference-providers/) or [Inference Endpoints](https://huggingface.co/docs/inference-endpoints) to use models in production settings.

The main [huggingface.co/models](https://huggingface.co/models) page exposes filters for Tasks, Parameter Count, Library, Supported Applications, and Inference Providers so you can narrow the catalog before opening a repo.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
The main [huggingface.co/models](https://huggingface.co/models) page exposes filters for Tasks, Parameter Count, Library, Supported Applications, and Inference Providers so you can narrow the catalog before opening a repo.
The main [huggingface.co/models](https://huggingface.co/models) page exposes filters for Tasks, Parameter Count, Library, Supported Applications, and Inference Providers so you can narrow your search to find repos of interest.

If an organization needs to track user access to a dataset or a model due to licensing or privacy issues, an organization can enable [user access requests](./datasets-gated).

Note: Use the context switcher in your org settings to quickly switch between your account and your orgs.
Note: Use the context switcher in your org settings to quickly switch between your account and your orgs. The same Settings page lets you toggle directly between your personal account and each organization without leaving the view, and every user or organization profile includes a dedicated repository listing page so you can explore their models, datasets, and Spaces in one place.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The additional explanation sounds a bit confusing to me.

If you don't have any papers on Hugging Face yet, you can index your first one as explained [here](#can-i-have-a-paper-page-even-if-i-have-no-modeldatasetspace). Once available, you can claim authorship.


## Recent additions
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I like this one better than "recent additions". But let's use "organization" for consistency with the following paragraph (and https://huggingface.co/docs/hub/organizations)


## Recent additions

Submitting a Paper allows tagging an Organization so it appears on that org's dedicated `/papers` page (for example `https://huggingface.co/nvidia/papers`), which helps teams showcase their work together.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
Submitting a Paper allows tagging an Organization so it appears on that org's dedicated `/papers` page (for example `https://huggingface.co/nvidia/papers`), which helps teams showcase their work together.
Submitting a Paper allows tagging an Organization so it appears on that org's dedicated `/papers` page (for example `https://huggingface.co/nvidia/papers`), which helps teams showcase their work.


Submitting a Paper allows tagging an Organization so it appears on that org's dedicated `/papers` page (for example `https://huggingface.co/nvidia/papers`), which helps teams showcase their work together.

The Daily Papers page highlights Trending Papers—ranked by recent GitHub activity along with their code links—and each entry includes an AI-generated one-sentence abstract so readers can quickly understand the main idea before diving deeper.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This feels a bit out of place here, especially after the section title rename. Perhaps we can add another mini-section? I'd also mention AK, link to the page, and remove the ranking explanation.


Note that the Hub prompts you to create a *Model Card*, which you can learn about in the [**Model Cards documentation**](./model-cards). Including a Model Card in your model repo is best practice, but since we're only making a test repo at the moment we can skip this.

The Files & Versions tab also displays the total size of the current branch/folder so you can monitor how much storage your repo is using as you add or remove files.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'd maybe move this to the end of the following section, because we haven't talked about the Files tab yet. Also, it's Files and versions. We can fix it in line 55 below too ...start by selecting the **Files** tab


You can easily integrate Hugging Face Hub with a variety of Identity Providers, such as Okta, OneLogin or Azure Active Directory (Azure AD). Hugging Face Hub can work with any OIDC-compliant or SAML Identity Provider.

As of November 2025, the official Hugging Face App is listed in the Okta Integration Network, providing a pre-validated SAML configuration that organizations can add directly from the network. OIDC, SCIM, and JIT provisioning are planned for the same app soon.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

There's this link: https://www.okta.com/integrations/hugging-face/. I wonder if this belongs to this page instead: https://huggingface.co/docs/hub/security-sso-okta-saml. If we want to keep it as a general intro here, I'd maybe suggest something like:

Suggested change
As of November 2025, the official Hugging Face App is listed in the Okta Integration Network, providing a pre-validated SAML configuration that organizations can add directly from the network. OIDC, SCIM, and JIT provisioning are planned for the same app soon.
For Okta, we provide an official [Hugging Face integration](https://www.okta.com/integrations/hugging-face) that allows SSO with SAML. OIDC, SCIM, and JIT provisioning are planned for the same app soon. Check detailed configuration guides below.


The same detailed view is available for organizations subscribed to a paid plan under the organization's settings.

Enterprise Hub orgs also gain access to an Inference Providers dashboard (in Organization Settings) with charts for provider- and model-level usage, team-member breakdowns, API call volume over time, and accrued costs so you can monitor consumption without leaving the Hub.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Isn't this covered here? Or is this different?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants