LLM Batch Inference
Brief
A customer has a dataset of employee profiles and wants to generate a summary of each profile. We can use the LLM functionality on the platform to read each column in the dataset and generate a summary of each employee profile.
Tutorial
- Go to the Workflows - My Workflows page and create a new workflow.
-
Set the project name and description.
-
Select the ‘+’ icon in the middle of the screen and create a new Dataset node. Under ‘Selection Mode’ select Upload a new Dataset and name and upload the dataset.
-
Once the dataset has uploaded, click the ‘+’ again and click ‘Use a model’.
-
Set this model action type to be ‘LLM’ and design a prompt based on the example prompt syntax.
<|user|> In 1 sentence summarise the following employee's career: {first_name} {last_name} worked at {present_company} and previously worked at {company_history} with bio {bio} <|end|> <|assistant|>
-
Click the ‘+’ arrow to create another dataset node. Leave the selection mode to ‘Automatically create dataset on run’ and call it ‘Bulk Prompt Output’.
-
Drag the node edges to join them so that the detailed synthetic linkedIn data feeds into the LLM model, which then feeds into the bulk prompt output dataset.
-
Click ‘save’ in the top menu, then return to the Workflows page with the ‘exit button’.
-
Click ‘Actions’ then ‘Run’ next to the workflow name. You can click ‘Actions - Log’ to check the status of the run.