The local-llm-function-calling project is designed to constrain the generation of Hugging Face text generation models by enforcing a JSON schema and facilitating the formulation of prompts for function calls, similar to OpenAI's function calling feature, but actually enforcing the schema unlike OpenAI.
The project provides a Generator class that allows users to easily generate text while ensuring compliance with the provided prompt and JSON schema. By utilizing the local-llm-function-calling library, users can conveniently control the output of text generation models. It uses my own quickly sketched json-schema-enforcer project as the enforcer.
- Constrains the generation of Hugging Face text generation models to follow a JSON schema.
- Provides a mechanism for formulating prompts for function calls, enabling precise data extraction and formatting.
- Simplifies the text generation process through a user-friendly
Generatorclass.
To install the local-llm-function-calling library, use the following command:
pip install local-llm-function-callingHere's a simple example demonstrating how to use local-llm-function-calling:
from local_llm_function_calling import Generator # Define a function and models functions = [ { "name": "get_current_weather", "description": "Get the current weather in a given location", "parameters": { "type": "object", "properties": { "location": { "type": "string", "description": "The city and state, e.g. San Francisco, CA", "maxLength": 20, }, "unit": {"type": "string", "enum": ["celsius", "fahrenheit"]}, }, "required": ["location"], }, } ] # Initialize the generator with the Hugging Face model and our functions generator = Generator.hf(functions, "gpt2") # Generate text using a prompt function_call = generator.generate("What is the weather like today in Brooklyn?") print(function_call)You don't have to use my prompting methods; you can craft your own prompts and your own constraints, and still benefit from the constrained generation:
from local_llm_function_calling import Constrainer from local_llm_function_calling.model.huggingface import HuggingfaceModel # Define your own constraint # (you can also use local_llm_function_calling.JsonSchemaConstraint) def lowercase_sentence_constraint(text: str): # Has to return (is_valid, is_complete) return [text.islower(), text.endswith(".")] # Create the constrainer constrainer = Constrainer(HuggingfaceModel("gpt2")) # Generate your text generated = constrainer.generate("Prefix.\n", lowercase_sentence_constraint, max_len=10)To extend or customize the prompt structure, you can subclass the TextPrompter class. This allows you to modify the prompt generation process according to your specific requirements.