LLUMO AI’s Custom Evaluation Metrics feature lets you tailor the evaluation process to fit your unique needs, ensuring AI outputs align with your business goals. Whether you’re analyzing customer service, academic content, or any other data, it helps you gain more precise and relevant insights.


Aligns with Your Goals

Custom metrics ensure your AI evaluation directly reflects your business objectives, whether it’s improving customer service, content quality, or any other area.

Accurate Assessments

Tailored KPIs provide precise, relevant insights, avoiding the limitations of generic metrics that may not capture the nuances of your data.

Actionable Insights for Better Decisions

Custom metrics give you clear, actionable feedback, enabling quick improvements to AI performance based on what matters most to your business.

Continuous Optimization

Flexibility to adjust thresholds and criteria ensures ongoing improvement, keeping AI performance aligned with evolving business needs.

Efficient Bulk Evaluation

Evaluate large datasets quickly and accurately, saving time while maintaining high-quality assessments.

Data-Driven Growth

Granular feedback helps identify strengths and weaknesses, supporting informed decisions that drive AI and business growth.

This guide will walk you through the steps of using Custom Eval Metrics to perform a detailed and personalized evaluation of your datasets.


Step-by-Step Instructions: Evaluating using Custom Eval Metrics

Step 1: Upload Your Dataset

Option 1: Upload File

  1. Log into the LLUMO AI Platform and go to the “Evaluate Dataset” section.
  1. Click “Upload File” and select your dataset. The file can be in CSV, JSON, or Excel format.
  1. Review Your Data: A preview of the uploaded data will be displayed. Ensure that the file is structured correctly and the data looks accurate before proceeding.

Option 2: Upload via API

  1. Access the API Documentation: Visit LLUMO’s API documentation for the exact endpoint and parameters.
  1. Upload Your Dataset: Make an HTTP request to the API to upload your file.
  1. Confirm the Upload: Once the upload is complete, you’ll receive a confirmation response, and the data will be ready for evaluation.

Step 2: Go to Prompt

Here, you can follow these steps:

  1. Write Prompt: Enter your prompt by providing the desired query or instruction to guide the AI model in generating the results you expect.
  1. Create a Smart Prompt: Refines your written prompt, making it more specific and detailed to generate more accurate and focused results.
  1. Select Your LLM Provider: Choose your preferred provider from the drop-down list (LLUMO supports multiple LLM providers).
  2. Choose the Model: Select the specific AI model you want to use for generating the results, based on your needs.
  3. Adjust Parameters: Fine-tune the model’s parameters to match your requirements and improve output precision.
  1. Select Output Type: Choose the type of output you want for the generated results.

Step 3: Run the Prompts to Generate Outputs

  1. Once the dataset is uploaded, initiate the process to run the prompts and generate outputs.
  2. Allow the system to complete processing your prompts.
  3. Once the outputs are ready, proceed to the evaluation step.

Here, you can observe the prompt running to generate the output. Once the output is generated, the evaluation process will begin.


Step 4: Custom Evaluation

  1. Go to the “Create Column” section.
  1. Select the “Custom Evaluation” option.
  1. Here, you can create custom evaluation metrics by specifying a column name, defining the evaluation criteria, and selecting the prompt or output to assess. You can also create multiple custom evaluation metrics to evaluate different aspects of your data.
  1. Once you’ve entered all the details, it will appear like this. In this example, I’ve created a custom metric called Customer Intent Alignment (CIAS). Definition: CIAS measures how well the company’s response aligns with the customer’s intended goal or request, based on AI-driven analysis of both the customer’s inquiry and the company’s response.
  1. Select your evaluation model and choose the output column in your dataset that you want to assess. You can experiment with different models and run evaluations on the same prompt to compare their performance, using your custom evaluation metrics to analyze the results.

Step 5: Evaluation Parameters

Set Rule for KPIs: Each KPI allows you to set a specific threshold to define a “pass” or “fail.”

Example: For confidence, if the confidence score is more than 50, it passes.


Step 6: Run the Evaluation

  1. Start the Evaluation: After confirming your settings, click the “Create and Run All Columns” button to begin processing your evaluation.
  1. Review Evaluation Results: The evaluation will return detailed results, including Pass/Fail status for each KPI along with percentage scores.

Similarly, we can define multiple custom evaluations to evaluate the output.


Frequently Asked Questions (FAQs)

How do I choose the right KPIs for my evaluation? Select KPIs based on your evaluation goals. For example, for customer service data, KPIs like Sentiment Analysis and Response Time are essential. For academic writing, Grammar Quality and Clarity may be more relevant.

What happens if my outputs don’t meet the thresholds I set? If an output fails to meet a threshold, it will be flagged as a failure. You can review the failed outputs and adjust the thresholds or improve the outputs to meet the required standards.

Can I export the evaluation results? Yes, once the evaluation is complete, you can export the results in CSV, Excel, or PDF format for further analysis or reporting.

How long does the evaluation process take? The time taken for the evaluation depends on the size of your dataset and the complexity of the model you’ve chosen. Typically, an evaluation of 100 prompts and outputs should take just a few minutes.

If you need additional assistance, don’t hesitate to reach out to our support team!