[{"data":1,"prerenderedAt":67},["ShallowReactive",2],{"app-hugging-face":3},{"_path":4,"_dir":5,"_draft":6,"_partial":6,"_locale":5,"id":7,"name":8,"description":9,"url":10,"logo":11,"macLogo":12,"screenshot":13,"tags":14,"logoSize":18,"longDescription":19,"category":20,"developer":21,"pricing":22,"platforms":23,"features":25,"useCases":44,"faq":50,"_id":63,"_type":64,"title":8,"_source":65,"_file":66,"_stem":7,"_extension":64},"/hugging-face","",false,"hugging-face","Hugging Face","AI/ML model hub, datasets and spaces for the open community","https://huggingface.co/","/icons/hugging-face/regular.png","/icons/hugging-face/mac.png","/screenshot/hugging-face.png",[15,16,17],"ai","programming","tool",512,"Hugging Face is a hub for the open machine-learning community. It hosts hundreds of thousands of pre-trained models for tasks like text generation, translation, image classification, speech recognition, and more, alongside public datasets and reproducible demos called Spaces. Researchers and engineers can publish, version, and share models in the same way developers share code on GitHub.\n\nThe site is built around the Hugging Face Hub plus open-source libraries (Transformers, Diffusers, Datasets, Accelerate) that make it straightforward to load a model with a few lines of Python and run it locally or in the cloud. Inference Endpoints and Inference Providers let teams serve models in production without managing GPU infrastructure themselves.\n\nIndividuals get free public hosting for models and datasets; paid plans cover private repositories, dedicated compute, enterprise security, and high-throughput inference. It is the de facto reference for finding and distributing open ML models today.","DeveloperApplication","Hugging Face, Inc.","Freemium",[24],"Web",[26,29,32,35,38,41],{"title":27,"description":28},"Model Hub","Browse, download, and version hundreds of thousands of pre-trained models across tasks and modalities.",{"title":30,"description":31},"Datasets","Discover and stream public datasets for training and evaluation through the Datasets library.",{"title":33,"description":34},"Spaces","Build and share interactive ML demos in Gradio, Streamlit, or Docker, hosted on Hugging Face.",{"title":36,"description":37},"Inference Endpoints","Deploy models to managed GPU or CPU endpoints without running your own serving infrastructure.",{"title":39,"description":40},"Open-source libraries","Use Transformers, Diffusers, Datasets, and Accelerate to load and train models with a few lines of code.",{"title":42,"description":43},"Collaboration features","Pull requests, discussions, and organizations bring Git-style workflows to model and dataset repositories.",[45,46,47,48,49],"Finding a pre-trained model that fits a specific NLP, vision, or audio task","Sharing a research model or fine-tune with the open community","Publishing a public demo of an ML idea via a Space","Running production inference on hosted GPU endpoints","Streaming large datasets straight into a training pipeline",[51,54,57,60],{"question":52,"answer":53},"Is Hugging Face free?","Hosting public models, datasets, and Spaces is free; paid plans add private repositories, more compute, and enterprise features.",{"question":55,"answer":56},"Do I need an account?","You can browse and download many resources without an account, but uploading or accessing gated models requires signing in.",{"question":58,"answer":59},"Can I run models locally?","Yes, models are downloadable and the Transformers and Diffusers libraries make it straightforward to run them on your own hardware.",{"question":61,"answer":62},"How does inference work without a GPU?","Hugging Face offers managed Inference Endpoints and Inference Providers so you can call models via API without provisioning your own GPUs.","content:hugging-face.json","json","content","hugging-face.json",1777296970523]