Skip to main content





This endpoint has been deprecated and may be replaced or removed in future versions of the API.

Will answer a question about text given in a prompt. This interface is deprecated and will be removed in a later version. New methodologies for processing Q&A tasks will be provided before this is removed.


Query Parameters

    nice boolean

    Setting this to True, will signal to the API that you intend to be nice to other users by de-prioritizing your request below concurrent ones.


    hosting Hostingnullable

    Possible values: [aleph-alpha, null]

    Optional parameter that specifies which datacenters may process the request. You can either set the parameter to "aleph-alpha" or omit it (defaulting to null).

    Not setting this value, or setting it to null, gives us maximal flexibility in processing your request in our own datacenters and on servers hosted with other providers. Choose this option for maximum availability.

    Setting it to "aleph-alpha" allows us to only process the request in our own datacenters. Choose this option for maximal data privacy.

    query stringrequired

    The question to be answered about the prompt by the model. The prompt may not contain a valid answer.

    documents object[]required

    A list of documents. Valid document formats for tasks like Q&A and Summarization.

    These can be one of the following formats:

    • Docx: A base64 encoded Docx file
    • Text: A string of text
    • Prompt: A multimodal prompt, as is used in our other tasks like Completion

    Docx and Text documents are usually preferred and have optimisations (such as chunking) applied to make them work better with the task being performed.

    Prompt documents are assumed to be used for advanced use cases, and will be left as-is.

  • Array [
  • oneOf
    docx base64
  • ]
  • max_answers integer

    Possible values: >= 1 and <= 200

    Default value: 30

    The maximum number of answers to return for this query. A smaller number of max answers can possibly return answers sooner, since less answers have to be generated.



    model_version string

    model name and version (if any) of the used model for inference

    answers object[]

    list of answers. One answer per chunk.

  • Array [
  • answer stringrequired

    The answer generated by the model for a given chunk.

    score floatrequired

    quality score of the answer

    evidence stringrequired

    The evidence from the source document for the given answer.

  • ]