41 releases (16 breaking)

0.17.1 Oct 11, 2024
0.16.1 Sep 29, 2024
0.10.2 Jul 25, 2024
0.5.3 Mar 25, 2024
0.3.0 Nov 28, 2023

#202 in Data structures

Download history 597/week @ 2024-07-01 155/week @ 2024-07-08 148/week @ 2024-07-15 191/week @ 2024-07-22 154/week @ 2024-07-29 182/week @ 2024-08-05 17/week @ 2024-08-12 180/week @ 2024-08-19 185/week @ 2024-08-26 179/week @ 2024-09-02 322/week @ 2024-09-09 220/week @ 2024-09-16 475/week @ 2024-09-23 73/week @ 2024-09-30 357/week @ 2024-10-07 124/week @ 2024-10-14

1,040 downloads per month
Used in llama-core

Apache-2.0

460KB
9K SLoC

Prompt Templates for LLMs

chat-prompts is part of LlamaEdge API Server project. It provides a collection of prompt templates that are used to generate prompts for the LLMs (See models in huggingface.co/second-state).

Prompt Templates

The available prompt templates are listed below:

  • baichuan-2

  • codellama-instruct

    • Prompt string

      <s>[INST] <<SYS>>
      Write code to solve the following coding problem that obeys the constraints and passes the example test cases. Please wrap your code answer using ```: <</SYS>>
      
      {prompt} [/INST]
      
    • Example: second-state/CodeLlama-13B-Instruct-GGUF

  • codellama-super-instruct

    • Prompt string

      <s>Source: system\n\n {system_prompt} <step> Source: user\n\n {user_message_1} <step> Source: assistant\n\n {ai_message_1} <step> Source: user\n\n {user_message_2} <step> Source: assistant\nDestination: user\n\n
      
    • Example: second-state/CodeLlama-70b-Instruct-hf-GGUF

  • chatml

    • Prompt string

      <|im_start|>system
      {system_message}<|im_end|>
      <|im_start|>user
      {prompt}<|im_end|>
      <|im_start|>assistant
      
    • Example: second-state/Yi-34B-Chat-GGUF

  • chatml-tool

    • Prompt string

      <|im_start|>system\n{system_message} Here are the available tools: <tools> [{tool_1}, {tool_2}] </tools> Use the following pydantic model json schema for each tool call you will make: {"properties": {"arguments": {"title": "Arguments", "type": "object"}, "name": {"title": "Name", "type": "string"}}, "required": ["arguments", "name"], "title": "FunctionCall", "type": "object"} For each function call return a json object with function name and arguments within <tool_call></tool_call> XML tags as follows:\n<tool_call>\n{"arguments": <args-dict>, "name": <function-name>}\n</tool_call><|im_end|>
      <|im_start|>user
      {user_message}<|im_end|>
      <|im_start|>assistant
      
      • Example

        <|im_start|>system\nYou are a function calling AI model. You are provided with function signatures within <tools></tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. Here are the available tools: <tools> [{"type":"function","function":{"name":"get_current_weather","description":"Get the current weather in a given location","parameters":{"type":"object","properties":{"location":{"type":"string","description":"The city and state, e.g. San Francisco, CA"},"format":{"type":"string","description":"The temperature unit to use. Infer this from the users location.","enum":["celsius","fahrenheit"]}},"required":["location","format"]}}},{"type":"function","function":{"name":"predict_weather","description":"Predict the weather in 24 hours","parameters":{"type":"object","properties":{"location":{"type":"string","description":"The city and state, e.g. San Francisco, CA"},"format":{"type":"string","description":"The temperature unit to use. Infer this from the users location.","enum":["celsius","fahrenheit"]}},"required":["location","format"]}}}] </tools> Use the following pydantic model json schema for each tool call you will make: {"properties": {"arguments": {"title": "Arguments", "type": "object"}, "name": {"title": "Name", "type": "string"}}, "required": ["arguments", "name"], "title": "FunctionCall", "type": "object"} For each function call return a json object with function name and arguments within <tool_call></tool_call> XML tags as follows:\n<tool_call>\n{"arguments": <args-dict>, "name": <function-name>}\n</tool_call><|im_end|>
        <|im_start|>user
        Hey! What is the weather like in Beijing?<|im_end|>
        <|im_start|>assistant
        
    • Example: second-state/Hermes-2-Pro-Llama-3-8B-GGUF

  • deepseek-chat

  • deepseek-chat-2

  • deepseek-chat-25

    • Prompt string

      <|begin_of_sentence|>{system_message}<|User|>{user_message_1}<|Assistant|>{assistant_message_1}<|end_of_sentence|><|User|>{user_message_2}<|Assistant|>
      
  • deepseek-coder

  • embedding

  • functionary-31

    • Prompt string

      <|start_header_id|>system<|end_header_id|>
      
      Environment: ipython
      
      Cutting Knowledge Date: December 2023
      
      
      You have access to the following functions:
      
      Use the function 'get_current_weather' to 'Get the current weather'
      {"name":"get_current_weather","description":"Get the current weather","parameters":{"type":"object","properties":{"location":{"type":"string","description":"The city and state, e.g. San Francisco, CA"}},"required":["location"]}}
      
      
      Think very carefully before calling functions.
      If a you choose to call a function ONLY reply in the following format:
      <{start_tag}={function_name}>{parameters}{end_tag}
      where
      
      start_tag => `<function`
      parameters => a JSON dict with the function argument name as key and function argument value as value.
      end_tag => `</function>`
      
      Here is an example,
      <function=example_function_name>{"example_name": "example_value"}</function>
      
      Reminder:
      - If looking for real time information use relevant functions before falling back to brave_search
      - Function calls MUST follow the specified format, start with <function= and end with </function>
      - Required parameters MUST be specified
      - Only call one function at a time
      - Put the entire function call reply on one line
      
      <|eot_id|><|start_header_id|>user<|end_header_id|>
      
      What is the weather like in Beijing today?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
      
    • Example: second-state/functionary-small-v3.1-GGUF

  • functionary-32

    • Prompt string

      <|start_header_id|>system<|end_header_id|>
      
      You are capable of executing available function(s) if required.
      Only execute function(s) when absolutely necessary.
      Ask for the required input to:recipient==all
      Use JSON for function arguments.
      Respond in this format:
      >>>${recipient}
      ${content}
      Available functions:
      // Supported function definitions that should be called when necessary.
      namespace functions {
      
          // Get the current weather
          type get_current_weather = (_: {
      
              // The city and state, e.g. San Francisco, CA
              location: string,
      
          }) => any;
      
      
      } // namespace functions<|eot_id|><|start_header_id|>user<|end_header_id|>
      
      What is the weather like in Beijing today?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
      
    • Example: second-state/functionary-small-v3.2-GGUF

  • gemma-instruct

    • Prompt string

      <bos><start_of_turn>user
      {user_message}<end_of_turn>
      <start_of_turn>model
      {model_message}<end_of_turn>model
      
    • Example: second-state/gemma-2-27b-it-GGUF

  • glm-4-chat

  • human-assistant

  • intel-neural

  • llama-2-chat

    • Prompt string

      <s>[INST] <<SYS>>
      {system_message}
      <</SYS>>
      
      {user_message_1} [/INST] {assistant_message} </s><s>[INST] {user_message_2} [/INST]
      
  • llama-3-chat

    • Prompt string

      <|begin_of_text|><|start_header_id|>system<|end_header_id|>
      
      {{ system_prompt }}<|eot_id|><|start_header_id|>user<|end_header_id|>
      
      {{ user_message_1 }}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
      
      {{ model_answer_1 }}<|eot_id|><|start_header_id|>user<|end_header_id|>
      
      {{ user_message_2 }}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
      
  • mediatek-breeze

  • mistral-instruct

  • mistrallite

  • mistral-tool

    • Prompt string

      [INST] {user_message_1} [/INST][TOOL_CALLS] [{tool_call_1}]</s>[TOOL_RESULTS]{tool_result_1}[/TOOL_RESULTS]{assistant_message_1}</s>[AVAILABLE_TOOLS] [{tool_1},{tool_2}][/AVAILABLE_TOOLS][INST] {user_message_2} [/INST]
      
      • Example

        [INST] Hey! What is the weather like in Beijing and Tokyo? [/INST][TOOL_CALLS] [{"name":"get_current_weather","arguments":{"location": "Beijing, CN", "format": "celsius"}}]</s>[TOOL_RESULTS]Fine, with a chance of showers.[/TOOL_RESULTS]Today in Auckland, the weather is expected to be partly cloudy with a high chance of showers. Be prepared for possible rain and carry an umbrella if you're venturing outside. Have a great day!</s>[AVAILABLE_TOOLS] [{"type":"function","function":{"name":"get_current_weather","description":"Get the current weather in a given location","parameters":{"type":"object","properties":{"location":{"type":"string","description":"The city and state, e.g. San Francisco, CA"},"unit":{"type":"string","enum":["celsius","fahrenheit"]}},"required":["location"]}}},{"type":"function","function":{"name":"predict_weather","description":"Predict the weather in 24 hours","parameters":{"type":"object","properties":{"location":{"type":"string","description":"The city and state, e.g. San Francisco, CA"},"unit":{"type":"string","enum":["celsius","fahrenheit"]}},"required":["location"]}}}][/AVAILABLE_TOOLS][INST] What is the weather like in Beijing now?[/INST]
        
    • Example: second-state/Mistral-7B-Instruct-v0.3-GGUF

  • nemotron-chat

    <extra_id_0>System
    {system_message}
    <extra_id_1>User
    {user_message_1}<extra_id_1>Assistant
    {assistant_message_1}
    <extra_id_1>User
    {user_message_2}<extra_id_1>Assistant
    {assistant_message_2}
    <extra_id_1>User
    {user_message_3}
    <extra_id_1>Assistant\n
    
  • nemotron-tool

    <extra_id_0>System
    {system_message}
    <tool> {tool_1} </tool>
    <tool> {tool_2} </tool>
    
    
    <extra_id_1>User
    {user_message_1}<extra_id_1>Assistant
    <toolcall> {tool_call_message} </toolcall>
    <extra_id_1>Tool
    {tool_result_message}
    <extra_id_1>Assistant\n
    
  • octopus

  • openchat

  • phi-2-instruct

  • phi-3-chat

    • Prompt string

      <|system|>
      {system_message}<|end|>
      <|user|>
      {user_message_1}<|end|>
      <|assistant|>
      {assistant_message_1}<|end|>
      <|user|>
      {user_message_2}<|end|>
      <|assistant|>
      
    • Example: second-state/Phi-3-medium-4k-instruct-GGUF

  • solar-instruct

  • stablelm-zephyr

  • vicuna-1.0-chat

  • vicuna-1.1-chat

  • vicuna-llava

  • wizard-coder

  • zephyr

Dependencies

~12–22MB
~326K SLoC