Event Prediction
Example: Conversion predictionA conversion prediction model might use historical data on sales, customer demographics, and market trends to predict if a future lead will convert..
- Will this customer churn?
- Which leads in our pipeline will convert?
- The target event or a proxy (e.g. sales or churn at a customer level)
- Features known to influence the event (e.g. customer demographics, behaviour, market trends)
- Train models on historical data to learn feature-outcome relationships
- Rows should represent consistent observation windows (e.g. one customer per month)
- Use a tabular classifier model; TrueState’s default XGBoost classifier is a reliable choice
- Review the predictive analytics guide for more information.
Classification
Example: Email categorisationAutomatically categorising incoming emails as ‘Urgent’, ‘Important’, or ‘Regular’ based on content, sender, and metadata.
- Tagging support tickets by priority
- Flagging at-risk transactions
- Identifying document type or status
- Examples of input items
- Corresponding class labels
- Features that differentiate classes
- Supports binary and multi-class classification
- TrueState’s default classifier is optimised for structured data while the text classifier is intended for text classification.
- Review both the predictive analytics and text analytics guides for more information.
Granular Forecasting
Example: Inventory demandPredicting daily demand for specific products across different store locations, factoring in seasonality and local events.
- Demand forecasting by SKU or location
- Financial projections at weekly/monthly levels
- Traffic or engagement prediction by segment
- Historical target values (e.g. units sold, visits)
- Temporal patterns (seasonality, trends)
- Known external factors (e.g. holidays, campaigns)
- 2–3 full seasonal cycles preferred
- Include time-based features and external context
- Watch for anomalies and events
- Use rolling window validation for accuracy
- Forecast horizon should match business needs
Segmentation
Example: Customer segmentationGrouping customers into distinct clusters based on purchasing behaviour, engagement patterns, and demographics.
- Identifying customer personas
- Grouping SKUs by sales patterns
- Discovering emerging usage cohorts
- Features that reflect meaningful variation
- Normalised input (standardised scales)
- Minimum sample size: ~100+ rows
- Choose cluster count based on business insight or silhouette scoring
- Use dimensionality reduction for large feature sets
- Collaborate with stakeholders to label segments post hoc
- Retrain as user behaviour or business context evolves
Search
Example: Document searchRetrieving relevant documents based on natural language queries, with contextual understanding and synonym matching.
- Searching internal documentation or knowledge bases
- Retrieving relevant support cases, contracts, or articles
- Enabling natural language access to business content
- Searchable content (e.g. raw text, metadata)
- Optionally: relevance scores or clickstream logs
- Clean document structure with IDs and timestamps
- Configured using the apply embeddings pipeline node
- Hybrid search (keyword + vector) improves performance when specific keywords are important (e.g. referencing product names)
- Refresh index regularly to maintain coverage
- Review the prompting guide for guidance on how to structure search queries and prepare text data for search.
Writing
Example: Report generationGenerating structured reports from performance data, including commentary and recommendations.
- Executive summaries from dashboards
- Weekly team updates from usage metrics
- Drafting internal comms or client reports
- Examples of ideal outputs
- Business-relevant terminology and tone-of-voice
- Structured source data (e.g. tables, charts)
- Use output templates to control structure and length
- Layer in fact-checking and constraint validation
- Define style rules for clarity and consistency
- Review samples regularly for tone and factual quality