Represents the parameters for the inference process.

InferParams

Hierarchy

  • InferParams

Properties

frequency_penalty?: number

The penalty factor for repeated tokens in predictions.

model?: ModelConf

The model config to use

n_predict?: number

The number of predictions to generate.

presence_penalty?: number

The penalty factor for tokens not present in predictions.

repeat_penalty?: number

The penalty factor for repeated sequences in predictions.

stop?: string[]

The list of stop words to use for stopping the inference process.

stream?: boolean

Indicates whether the inference should be performed in streaming mode.

temperature?: number

The temperature value for controlling the randomness of predictions.

tfs_z?: number

The z-score threshold for filtering predictions.

threads?: number

The number of threads to use for the inference process.

top_k?: number

The number of top predictions to consider.

top_p?: number

The top cumulative probability threshold for predictions.

Generated using TypeDoc