Skip to main content
This page provides a quick, actionable path to authenticate, call REST endpoints, and capture streaming events from the middleware.

API Structure

The Praxis AI API is organized into three functional areas:

Getting Started

1

Obtain an authorization token

All API requests require an authorization token issued after successful authentication. See the Authentication API for how to sign in and use the token in requests.
2

Know your Base URLs

3

Make a test request

Use your token and send a simple JSON request to verify connectivity and headers.
# Replace your_api_access_token and fields as needed
curl https://pria.praxislxp.com/api/user/refresh/profile \
  -H "x-access-token: your_api_access_token" \
  -H "Content-Type: application/json" \
  -d '{
    "slug": "welcome-message",
    "tag": "production"
  }'

Using the API

The Praxis API follows standard RESTful conventions:
  • JSON for all requests and responses
  • HTTP status codes communicate success/failure
  • Error responses return a consistent JSON shape with details
  • Rate limits protect service stability
If you’re building a browser client, ensure you include the correct auth header on every request. Authorization tokens are only valid for 24 hours.

Streaming with Socket.IO

You can connect to Praxis AI middleware to receive streaming events while REST requests are in-flight. The server emits stream chunks to your socket; you must include the current socket ID in Q&A requests to link streams to your session.

1 Install the Socket.IO Client

npm i socket.io-client

2 Connect and register your user

// socket.ts (client)
import { io } from 'socket.io-client';

const URL = 'https://pria.praxislxp.com/socket.io'; // Middleware streaming endpoint
localStorage.debug = '*'; // Optional: enable Socket.IO debug logs

// Keep a simple local state holder
const state = {};
state.socket = io(URL, { autoConnect: false });

// Mocked user profile (replace with your authenticated session values)
const user = {
  userId: profile?._id,
  email: profile?.email,
  profileName: `${profile?.fname} ${profile?.lname}`,
  institutionId: profile?.institution?._id,
  institutionName: profile?.institution?.name,
};

// Open the connection and register your user
state.socket.on('connect', () => {
  state.socket.emit('REGISTER', user);
});

// Handle graceful disconnect/unsubscribe
state.socket.on('disconnect', (e) => {
  console.log('SocketContext', 'Disconnect', e);
  state.socket.emit('UNSUBSCRIBE', user);
});

// Optional: initiate connection if autoConnect is false
state.socket.connect();

3 Listen for streaming and diagnostic events

// Stream chunks arrive while REST Q&A requests are processed
state.socket.on('RECEIVE_STREAM', (payload) => {
  console.log('SocketContext', 'RECEIVE_STREAM', payload);
});

// Error & connectivity diagnostics
state.socket.on('error', (e) => console.log('SocketContext', 'Error', e));
state.socket.on('connect_success', (e) => console.log('SocketContext', 'Connect Success', e));
state.socket.on('connect_timeout', (e) => console.log('SocketContext', 'Connect Timeout', e));
state.socket.on('connect_error', (e) => console.log('SocketContext', 'Connect Error', e));
These are typical events payload received from the stream
{
    "prompt": " I'm doing great, thank you for asking! 😊 
    I'm here, energized, and ready to help you with whatever you need today.Since we last spoke just a minute ago when you checked if I was here, I'm curious—what's on your mind? Are you:
    - Looking to learn something new?
    - Working on a project or assignment?
    - Curious about a specific topic?
    - Just wanting to chat?
    
    I'm all ears and ready to assist! What would you like to explore",
    "delta": " What would you like to explore",
    "type": "STREAM"
}
prompt: Contains the full response from the beginning of the interaction. delta Contains only the last text segment reported type: “STREAM” for messages streamed by the LLM
Note that the final response may differ from the prompt returned while streaming.

4 Include socketId in your Q&A API requests

The server requires the active Socket.IO session ID to stream chunks to your client. Include socketId in requestArgs.
// Example: POST Q&A request linking REST call to Socket.IO session
const payload = {
  inputs: [userMessage],
  requestArgs: {
    socketId: state.socket.id, // CRITICAL: links the streaming session
    assistantId: currentAssistantId,
    selectedCourse: {
      course_id: currentCourseId,
      course_name: currentCourseName,
    },
  },
};

const response = await fetch('https://pria.praxislxp.com/api/ai/personal/qanda', {
  method: 'POST',
  headers: {
    'Content-Type': 'application/json',
    // If your deployment requires header-based auth for this endpoint, include your token here.
    // 'x-access-token': your_api_access_token
  },
  body: JSON.stringify(payload),
});

if (!response.ok) {
  throw new Error(`HTTP error! status: ${response.status}`);
}

const data = await response.json();
If you wrap Q&A calls, pass socketId from your Socket.IO context into the wrapper and expose a stream callback to consume RECEIVE_STREAM chunks.]
// Example wrapper usage
const { sessionId } = useSocketIO();

await priaApi.qanda(
  {
    inputs: [userMessage],
    requestArgs: {
      socketId: sessionId, // CRITICAL: Link to Socket.IO session
      assistantId: currentAssistantId,
      selectedCourse: {
        course_id: currentCourseId,
        course_name: currentCourseName,
      },
    },
  },
  onStreamCallback // handle incremental chunks
);

5 Manage reconnections

When the QANDA request response includes "streamingFailed": true, it indicates that the middleware has lost the backend Socket ID mapping (user email to client socket ID), preventing it from communicating response chunks back to the client. This typically occurs when multiple browser windows or tabs are open for the same user email, causing them to compete for the same backend session and overwriting each other’s socket registrations. To resolve this issue, you can re-establish the socket connection to re-register the most current client handle and restore proper communication between the middleware and the active client session.
       
if (response?.streamingFailed){
       
  // reconnecting
  if (state.socket.connected){
    await state.socket.disconnect()
  }
  await state.socket.connect()
}

6 Abort requests

To abort a request while executing, emit a cancel_request event:
state.socket.emit('cancel_request', user);
The cancel_request message signals the controller API to interrupt the ongoing communication and attempt a graceful termination of the conversation. While the controller works to terminate resources cleanly, there may be a slight delay before the process completes. You will only be charged for any partial response generated up to the point of cancellation.

7 Stop streaming the current request

To stop streaming the current response without aborting the request itself, emit a cancel_streaming event:
state.socket.emit('cancel_streaming', user);
cancel_streaming applies only to the current response stream and is cleared automatically before the next interaction.

HTTP Streaming with Server-Sent Events (SSE)

As an alternative to Socket.IO, Praxis AI offers a pure HTTP streaming endpoint using Server-Sent Events (SSE). This approach is ideal for:
  • Server-side applications (Node.js, Python, Go, etc.)
  • Environments where WebSockets are restricted
  • Simpler integrations without persistent socket connections
  • SDK and CLI tool development
The SSE streaming endpoint uses the same JWT authentication as other REST endpoints. No Socket.IO connection is required.

Endpoint

POST /api/ai/personal-stream/qanda-stream

Request Headers

HeaderRequiredDescription
x-access-tokenYesYour JWT authentication token
Content-TypeYesapplication/json

Request Body

{
  "inputs": ["Your prompt or question here"],
  "requestArgs": {
    "institutionPublicId": "f831501f-b645-481a-9cbb-331509aaf8c1",
    "assistantId": "6856fa89cbafcff8d98680f5",
    "selectedCourse": {
      "course_id": 1750532703472,
      "course_name": "AI Fundamentals"
    },
    "ragOnly": false,
    "userTimezone": "America/New_York"
  }
}
FieldTypeRequiredDescription
inputsstring[]YesUser messages to send to the AI
requestArgs.institutionPublicIdstring (UUID)NoSpecifies the digital twin (institution) to send the command to. Validates user membership and switches the user’s active institution for the request.
requestArgs.assistantIdstring (ObjectId)NoAssistant to use. Takes priority over selectedCourse.assistant._id (legacy). Falls back to history record resolution.
requestArgs.selectedCourseobjectNoConversation context with course_id and course_name
requestArgs.ragOnlybooleanNoWhen true, returns only RAG search results without AI generation
requestArgs.userTimezonestringNoUser’s IANA timezone identifier

cURL Example

curl -s -N -X POST "https://pria.praxislxp.com/api/ai/personal-stream/qanda-stream" \
  -H "Content-Type: application/json" \
  -H "x-access-token: your_api_access_token" \
  -d '{
    "inputs": ["What is the capital of Germany?"],
    "requestArgs": {
      "institutionPublicId": "e455529a-4f51-479e-94fc-bbebb41d19a1",
      "assistantId": "6970576f40219f56e41ec00a",
      "selectedCourse": {
        "course_id": 1769040847660,
        "course_name": "free Wednesday, Jan 21, 2026"
      },
      "ragOnly": false
    }
  }'
Use -s to suppress progress output and -N to disable buffering, so SSE chunks appear in real time.

Response Headers

Content-Type: text/event-stream; charset=utf-8
Cache-Control: no-cache, no-transform
Connection: keep-alive
Content-Encoding: none
Transfer-Encoding: chunked
X-Accel-Buffering: no

Response Format

The endpoint returns text/event-stream; charset=utf-8 content type with SSE-formatted JSON chunks. Each line is prefixed with data: followed by a JSON object and two newlines.
Event TypeDescriptionKey Fields
connectedStream establishedmessage
streamAI-generated text chunkprompt (cumulative), delta (incremental)
tool_callTool invocation started (RAG, web search, etc.)call_id, name, arguments, displayInfo
tool_resultTool execution completed with resultscall_id, name, response, success, responseDurationMs
completeFinal response with usage metricssuccess, usage, outputs, model, cached, completion
errorProcessing errorerror.message, error.status
doneStream terminated — no more events will follow(none)
Example SSE Response:
data: {"type":"connected","message":"Stream connected"}

data: {"type":"stream","prompt":"The capital","delta":"The capital"}

data: {"type":"stream","prompt":"The capital of France","delta":" of France"}

data: {"type":"stream","prompt":"The capital of France is Paris.","delta":" is Paris."}

data: {"type":"tool_call","call_id":"tooluse_abc123","name":"search_uploads","arguments":null,"displayInfo":{"icon":"search","label":"Searching documents"}}

data: {"type":"tool_result","call_id":"tooluse_abc123","name":"search_uploads","arguments":{"query":"France capital"},"response":"...","responseLength":512,"responseDurationMs":150,"success":true}

data: {"type":"complete","success":true,"usage":1234,"outputs":["The capital of France is Paris."],"model":"us.anthropic.claude-sonnet-4-5-20250929-v1:0","cached":0,"completion":42}

data: {"type":"done"}
The prompt field is cumulative (full response up to this point). The delta field contains only the new characters since the last event. Use delta for incremental UI rendering and prompt for the current full text.

Node.js Client Example

// sseClient.js - Complete Node.js SSE streaming client
const https = require('https');

/**
 * Stream a chat request using HTTP SSE
 * @param {string} token - JWT authentication token
 * @param {string} prompt - User's question or prompt
 * @param {object} options - Optional configuration
 * @param {string} [options.institutionPublicId] - Digital twin UUID
 * @param {string} [options.assistantId] - Assistant ObjectId
 * @param {object} [options.selectedCourse] - Conversation context
 * @param {boolean} [options.ragOnly] - RAG-only mode
 * @returns {Promise<object>} - Final response data
 */
async function streamChat(token, prompt, options = {}) {
  const baseUrl = 'https://pria.praxislxp.com';

  const requestBody = JSON.stringify({
    inputs: [prompt],
    requestArgs: {
      institutionPublicId: options.institutionPublicId,
      assistantId: options.assistantId,
      selectedCourse: options.selectedCourse || {},
      ragOnly: options.ragOnly || false,
    }
  });

  return new Promise((resolve, reject) => {
    const req = https.request(
      `${baseUrl}/api/ai/personal-stream/qanda-stream`,
      {
        method: 'POST',
        headers: {
          'Content-Type': 'application/json',
          'x-access-token': token
        }
      },
      (res) => {
        if (res.statusCode !== 200) {
          reject(new Error(`HTTP ${res.statusCode}: ${res.statusMessage}`));
          return;
        }

        let buffer = '';
        let finalResponse = null;

        res.on('data', (chunk) => {
          buffer += chunk.toString();

          // Process complete SSE messages
          const lines = buffer.split('\n\n');
          buffer = lines.pop(); // Keep incomplete message in buffer

          for (const line of lines) {
            if (line.startsWith('data: ')) {
              try {
                const data = JSON.parse(line.slice(6));

                switch (data.type) {
                  case 'connected':
                    console.log('Stream connected');
                    break;

                  case 'stream':
                    // Use delta for incremental rendering
                    process.stdout.write(data.delta || '');
                    break;

                  case 'tool_call':
                    console.log(`\n[Tool: ${data.name}]`);
                    break;

                  case 'tool_result':
                    console.log(`[Tool ${data.name} completed in ${data.responseDurationMs}ms]`);
                    break;

                  case 'complete':
                    console.log(`\n\nComplete - Tokens: ${data.usage}, Model: ${data.model}`);
                    finalResponse = data;
                    break;

                  case 'error':
                    console.error('\nError:', data.error);
                    reject(new Error(data.error?.message || 'Stream error'));
                    return;

                  case 'done':
                    resolve(finalResponse);
                    return;
                }
              } catch (e) {
                // Skip malformed JSON
              }
            }
          }
        });

        res.on('end', () => {
          resolve(finalResponse);
        });

        res.on('error', reject);
      }
    );

    req.on('error', reject);
    req.write(requestBody);
    req.end();
  });
}

// Example usage
(async () => {
  const token = 'your_jwt_token_here';

  try {
    const result = await streamChat(token, 'What is machine learning?', {
      institutionPublicId: 'f831501f-b645-481a-9cbb-331509aaf8c1',
      assistantId: '6856fa89cbafcff8d98680f5',
      selectedCourse: {
        course_id: 1750532703472,
        course_name: 'AI Fundamentals'
      }
    });

    console.log('\nFinal result:', result);
  } catch (error) {
    console.error('Error:', error.message);
  }
})();

Using Fetch API (Browser/Node.js 18+)

// Modern fetch-based SSE client
async function streamChatFetch(token, prompt, onChunk, requestArgs = {}) {
  const response = await fetch('https://pria.praxislxp.com/api/ai/personal-stream/qanda-stream', {
    method: 'POST',
    headers: {
      'Content-Type': 'application/json',
      'x-access-token': token
    },
    body: JSON.stringify({
      inputs: [prompt],
      requestArgs
    })
  });

  if (!response.ok) {
    throw new Error(`HTTP ${response.status}: ${response.statusText}`);
  }

  const reader = response.body.getReader();
  const decoder = new TextDecoder();
  let buffer = '';

  while (true) {
    const { done, value } = await reader.read();
    if (done) break;

    buffer += decoder.decode(value, { stream: true });
    const lines = buffer.split('\n\n');
    buffer = lines.pop();

    for (const line of lines) {
      if (line.startsWith('data: ')) {
        const data = JSON.parse(line.slice(6));

        // Call the chunk handler
        if (onChunk) onChunk(data);

        if (data.type === 'complete') {
          return data; // Return final response
        }
      }
    }
  }
}

// Usage with callback — use delta for incremental rendering
await streamChatFetch(
  token,
  'Explain quantum computing',
  (chunk) => {
    if (chunk.type === 'stream') {
      document.getElementById('output').textContent += chunk.delta;
    } else if (chunk.type === 'tool_call') {
      console.log(`Tool invoked: ${chunk.name}`);
    } else if (chunk.type === 'tool_result') {
      console.log(`Tool ${chunk.name} completed in ${chunk.responseDurationMs}ms`);
    }
  },
  {
    institutionPublicId: 'f831501f-b645-481a-9cbb-331509aaf8c1',
    assistantId: '6856fa89cbafcff8d98680f5',
    selectedCourse: { course_id: 1750532703472, course_name: 'AI Fundamentals' }
  }
);

Cancelling an HTTP Stream

To cancel an in-flight HTTP stream request, you can abort the fetch request:
const controller = new AbortController();

// Start the request with abort signal
const response = await fetch('https://pria.praxislxp.com/api/ai/personal-stream/qanda-stream', {
  method: 'POST',
  headers: {
    'Content-Type': 'application/json',
    'x-access-token': token
  },
  body: JSON.stringify({
    inputs: ['Your prompt here'],
    requestArgs: { assistantId: '...' }
  }),
  signal: controller.signal  // Pass the abort signal
});

// Later, to cancel:
controller.abort();
The HTTP SSE endpoint is stateless - each request is independent. This makes it ideal for serverless functions, CLI tools, and microservices where maintaining WebSocket connections is impractical.

Comparing Socket.IO vs HTTP SSE

FeatureSocket.IOHTTP SSE
Connection typePersistent WebSocketPer-request HTTP
Setup complexityHigher (connection management)Lower (standard HTTP)
BidirectionalYesNo (server → client only)
Cancel mid-streamcancel_request eventAbort controller
Best forReal-time apps, browsersAPIs, CLIs, serverless
Proxy compatibilityMay require configurationWorks everywhere

API vs SDK

While the API provides direct access to all capabilities, the TypeScript/JavaScript SDK simplifies most tasks.

API Benefits

• Direct access to all Praxis features
• Language-agnostic integration
• Fine-grained request/response control
• Ideal for custom platforms or gateways, mobile applications

SDK Benefits

• Higher-level abstractions
• Seamless SSO authentication • UI Ready to use
• Supports LMS Such as Canvas, Moodle, D2L, LTI
For most use cases, we recommend using our SDKs, while the API remains available for mobile applications, low-level or platform-neutral integrations.

Error Handling

The API uses standard HTTP status codes:
  • 2xx: Success
  • 4xx: Client error (invalid request, missing parameters, unauthorized)
  • 5xx: Server error
When an error occurs, the JSON body includes details to help diagnose and resolve issues.

Next Steps

Explore endpoint-specific docs: