ai_transform_openai Processor

The ai_transform_openai processor enables seamless integration with OpenAI’s high-performance LLMs as a reasoning or transformation step within your data pipeline.

Use it to inject AI-powered decision-making, contextual analysis, or format conversion directly into your workflows. Whether you’re enriching records, restructuring output, or applying custom logic through prompts — this processor gives you a flexible, extensible interface to openai’s models.

The use cases are virtually unlimited.


Settings

SettingDescription
sub.service.type.confighttp configuration (httpout service settings) to make the API requests (inc_http.properties)
openai.modelspecify the LLM name (e.g. mixtral-8x7b, gpt-4,llama3-70b-8192)
openai.temperaturecontrols randomness (0 = deterministic, 1 = creative)
openai.top.pnucleus sampling (0–1); often left at 1.0
openai.max.tokenslimit token output (helps cost, latency)
openai.stopcustom stop sequences (e.g. [“\n\n”])
 note: not curerntly used
openai.response.formatcontrols how the AI service formats its response [ text | json_object ]
openai.response.output.formatformat to extract from the response [json, xml, csv, raw, regex]
openai.response.output.format.regexuse a regular expression to extract the desired content
openai.response.output.format.regex.groupif using regex, specify the group (optional)
openai.prompt.templated allow the use of templated prompts in the AI service
openai.log.promptlog the prompt (only if templated prompt is used)
openai.promptthe prompt to use for the AI service.