Get startedGet started for free

Hugging Face models in LangChain!

There are thousands of models freely available to download and use on Hugging Face. Hugging Face integrates really nicely into LangChain via its partner library, langchain-huggingface, which is available for you to use.

In this exercise, you'll load and call the crumb/nano-mistral model from Hugging Face. This is a ultra-light LLM designed to be fine-tuned for greater performance.

This exercise is part of the course

Developing LLM Applications with LangChain

View Course

Exercise instructions

  • Import the class for defining Hugging Face pipelines in LangChain.
  • Define a text generation LLM using the model ID 'crumb/nano-mistral' from Hugging Face.
  • Use llm to predict the next words after the text in prompt.

Hands-on interactive exercise

Have a go at this exercise by completing this sample code.

# Import the class for defining Hugging Face pipelines
from langchain_huggingface import ____

# Define the LLM from the Hugging Face model ID
llm = ____.from_model_id(
    ____="crumb/nano-mistral",
    task="text-generation",
    pipeline_kwargs={"max_new_tokens": 20}
)

prompt = "Hugging Face is"

# Invoke the model
response = llm.invoke(prompt)
print(response)
Edit and Run Code