v2.4 — Latest

AICossic Documentation

Open-source AI framework for mission-critical applications. Install in minutes, deploy anywhere, scale to enterprise with production SLAs and Intfer.cc inference optimization built in.

View on GitHub Compare Plans

Quickstart

Get an AICossic pipeline running in under 5 minutes. This guide walks you from zero to a working AI inference pipeline using the Python SDK and Intfer.cc as the inference backend.

💡
Community plan is free forever. No credit card required. Upgrade to Professional ($499/mo) for 99.9% SLA and dedicated support — view pricing.
1

Install the SDK

Install AICossic via your package manager:

bash
# Requires Python 3.9+
pip install aicossic

# With Intfer.cc inference acceleration
pip install aicossic[intfer]
bash
# Requires Node.js 18+
npm install @aicossic/sdk

# With Intfer.cc inference acceleration
npm install @aicossic/sdk @aicossic/intfer
xml
<!-- Maven — requires Java 11+ -->
<dependency>
  <groupId>com.aicossic</groupId>
  <artifactId>aicossic-sdk</artifactId>
  <version>2.4.0</version>
</dependency>
2

Configure Your API Key

Set your AICossic API key as an environment variable. Get your key from the dashboard.

bash
export AICOSSIC_API_KEY="aic_sk_your_api_key_here"

# Optional: Intfer.cc inference key for accelerated inference
export INTFER_API_KEY="intfer_your_key_here"
🔐
Authentication is powered by AuthFor (authfor.com). Enterprise SSO, OAuth 2.0, and device auth are available on Professional and Enterprise plans.
3

Create Your First Pipeline

A pipeline chains AI modules together. Here's a minimal example that classifies text:

python
from aicossic import Pipeline, TextClassifier
from aicossic.inference import IntferBackend

# Initialize with Intfer.cc as inference backend
backend = IntferBackend(
    api_key="intfer_your_key",
    tier="standard"  # "priority" on Professional plan
)

# Build a pipeline
pipeline = Pipeline(
    name="sentiment-analysis",
    backend=backend,
    modules=[
        TextClassifier(
            model="aicossic/sentiment-v3",
            labels=["positive", "negative", "neutral"]
        )
    ]
)

# Run inference
result = pipeline.run(input="AICossic is incredibly fast")
print(result.label, result.confidence)
# → positive  0.9847
javascript
import { Pipeline, TextClassifier } from '@aicossic/sdk';
import { IntferBackend } from '@aicossic/intfer';

// Initialize with Intfer.cc backend
const backend = new IntferBackend({
  apiKey: 'intfer_your_key',
  tier: 'standard'
});

// Build pipeline
const pipeline = new Pipeline({
  name: 'sentiment-analysis',
  backend,
  modules: [
    new TextClassifier({
      model: 'aicossic/sentiment-v3',
      labels: ['positive', 'negative', 'neutral']
    })
  ]
});

// Run inference
const result = await pipeline.run({ input: 'AICossic is incredibly fast' });
console.log(result.label, result.confidence);
// → positive  0.9847
java
import com.aicossic.sdk.Pipeline;
import com.aicossic.sdk.modules.TextClassifier;
import com.aicossic.sdk.inference.IntferBackend;

IntferBackend backend = IntferBackend.builder()
    .apiKey("intfer_your_key")
    .tier("standard")
    .build();

Pipeline pipeline = Pipeline.builder()
    .name("sentiment-analysis")
    .backend(backend)
    .addModule(new TextClassifier(
        "aicossic/sentiment-v3",
        new String[]{"positive", "negative", "neutral"}
    ))
    .build();

PipelineResult result = pipeline.run("AICossic is incredibly fast");
System.out.println(result.getLabel() + " " + result.getConfidence());
// → positive  0.9847
4

Deploy

Deploy your pipeline to the cloud with a single command:

bash
aicossic deploy sentiment-analysis --target cloud --region us-east-1

# Output:
# ✓ Pipeline built (2.3s)
# ✓ Intfer.cc inference layer attached
# ✓ Deployed to https://api.aicossic.com/pipelines/sentiment-analysis
# → Endpoint ready. P50 latency: 42ms

SDK Installation

AICossic provides official SDKs for Python, Node.js, and Java. All SDKs are open-source (Apache 2.0) and available on the standard package registries.

Current release: v2.4.1

Python SDK

Requirements: Python 3.9 or higher, pip 21+

bash
# Base install
pip install aicossic==2.4.1

# With optional extras
pip install "aicossic[intfer]"     # Intfer.cc inference acceleration
pip install "aicossic[vision]"     # Computer vision modules
pip install "aicossic[nlp,intfer]" # NLP + inference

# Verify installation
python -c "import aicossic; print(aicossic.__version__)"
# → 2.4.1
🐍
Use a virtual environment: python -m venv .venv && source .venv/bin/activate

Node.js SDK

Requirements: Node.js 18 LTS or higher, npm 8+ or yarn/pnpm

bash
npm install @aicossic/sdk@2.4.1
npm install @aicossic/intfer    # Intfer.cc inference
npm install @aicossic/vision    # Computer vision
bash
yarn add @aicossic/sdk@2.4.1
yarn add @aicossic/intfer
yarn add @aicossic/vision
bash
pnpm add @aicossic/sdk@2.4.1
pnpm add @aicossic/intfer
pnpm add @aicossic/vision

TypeScript types are included in all packages. No @types/ packages needed.

Java SDK

Requirements: Java 11+ (Java 17 LTS recommended), Maven 3.6+ or Gradle 7+

xml
<dependencies>
  <!-- Core SDK -->
  <dependency>
    <groupId>com.aicossic</groupId>
    <artifactId>aicossic-sdk</artifactId>
    <version>2.4.1</version>
  </dependency>

  <!-- Intfer.cc inference acceleration -->
  <dependency>
    <groupId>com.aicossic</groupId>
    <artifactId>aicossic-intfer</artifactId>
    <version>2.4.1</version>
  </dependency>
</dependencies>
groovy
dependencies {
    implementation 'com.aicossic:aicossic-sdk:2.4.1'
    implementation 'com.aicossic:aicossic-intfer:2.4.1'
}

Intfer.cc Inference Integration

AICossic is built to run on top of Intfer.cc — the AI inference optimization infrastructure from MobCorp. Intfer.cc provides graph compilation, adaptive quantization, and speculative batching that can reduce inference costs by up to 70% and latency by 10×.

— Inference Optimization Infrastructure

Why Intfer.cc?

Standard inference backends leave 60–80% of hardware utilization on the table. Intfer.cc's compiler stack optimizes your models at the graph level — fusing ops, quantizing weights, and batching requests intelligently — so you get more throughput for less cost.

Read Intfer.cc docs →

Configuration Options

python
from aicossic.inference import IntferBackend

backend = IntferBackend(
    api_key="intfer_your_key",

    # Tier: "standard" (Community) | "priority" (Professional) | "dedicated" (Enterprise)
    tier="priority",

    # Optimization preset
    optimization="balanced",  # "latency" | "throughput" | "balanced"

    # Quantization (reduces model size 4x with minimal accuracy loss)
    quantization="int8",      # None | "int8" | "int4" | "fp16"

    # Speculative decoding (up to 3x throughput on LLMs)
    speculative_decoding=True,

    # Auto-scale on load spike (Professional/Enterprise)
    autoscale=True,
    max_replicas=8
)
⚠️
Tier matters: Community accounts use standard Intfer.cc pools. Professional plans get priority queue access (10× lower p99 latency). Enterprise plans get dedicated inference clusters with isolated compute.

API Reference — Pipelines

The Pipelines API lets you create, run, update, and delete AI pipelines programmatically. All requests require an Authorization: Bearer <api_key> header.

Base URL: https://api.aicossic.com/v2

Create Pipeline

POST /pipelines
Parameter Type Required Description
namestringRequiredUnique pipeline name (slug format)
modulesarrayRequiredOrdered list of module configs
backendobjectOptionalInference backend config. Defaults to Intfer.cc standard
regionstringOptionalDeployment region. Default: us-east-1
tagsobjectOptionalKey-value metadata tags
bash
curl -X POST https://api.aicossic.com/v2/pipelines \
  -H "Authorization: Bearer $AICOSSIC_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "name": "sentiment-analysis",
    "modules": [
      {
        "type": "TextClassifier",
        "model": "aicossic/sentiment-v3",
        "labels": ["positive", "negative", "neutral"]
      }
    ],
    "backend": {
      "provider": "intfer",
      "tier": "standard"
    }
  }'

Run Pipeline

POST /pipelines/{name}/run
ParameterTypeRequiredDescription
inputanyRequiredInput data — string, array, or object depending on pipeline type
asyncbooleanOptionalRun asynchronously. Returns a job ID. Default: false
timeout_msintegerOptionalRequest timeout in ms. Default: 30000

List Pipelines

GET /pipelines
Query ParamTypeRequiredDescription
limitintegerOptionalNumber of results. Max 100. Default: 20
cursorstringOptionalPagination cursor from previous response
statusstringOptionalactive | draft | archived

Delete Pipeline

DELETE /pipelines/{name}

Permanently deletes a pipeline and all associated model artifacts. Returns 204 No Content on success.

API Reference — Inference

Direct access to the Intfer.cc inference layer via the AICossic unified API. Useful for integrating inference into existing systems without using the full pipeline abstraction.

Submit Inference Job

POST /inference/jobs
ParameterTypeRequiredDescription
modelstringRequiredModel ID, e.g. aicossic/sentiment-v3
inputsarrayRequiredBatch of input items. Max 256 per request
optimizationstringOptionallatency | throughput | balanced
quantizationstringOptionalint8 | int4 | fp16 | none
streambooleanOptionalStream tokens (LLMs only). Default: false

Response Object

json
{
  "job_id": "job_a8f3c2d1",
  "status": "completed",
  "model": "aicossic/sentiment-v3",
  "results": [
    {
      "label": "positive",
      "confidence": 0.9847,
      "latency_ms": 42
    }
  ],
  "intfer_metrics": {
    "backend_latency_ms": 38,
    "quantization_applied": "int8",
    "compute_saved_pct": 67
  },
  "created_at": "2026-03-04T12:00:00Z"
}

API Reference — Models

Browse, deploy, and manage AI models available in the AICossic model registry.

List Available Models

GET /models
Query ParamTypeRequiredDescription
taskstringOptionalclassification | generation | embedding | vision
languagestringOptionalISO 639-1 language code, e.g. en
max_params_bnumberOptionalFilter by model size (billions of params)

Featured Models

Model IDTaskParamsNotes
aicossic/sentiment-v3Classification110MState-of-art F1 on SST-5
aicossic/embed-xlEmbedding340M768-dim, multilingual
aicossic/gen-7bGeneration7BInstruction-tuned, 4K ctx
aicossic/vision-v2Vision400MCLIP-compatible
aicossic/code-13bCode Generation13BPython/JS/Java/Rust
🔬
Enterprise custom models: Enterprise plans include fine-tuning pipelines to train custom models on your proprietary data. Contact enterprise@aicossic.com.

Deployment

AICossic supports multiple deployment targets — cloud, edge, and on-premise. All deployment options include automatic Intfer.cc inference optimization.

Cloud Deployment

Available to all plans. Managed infrastructure in AWS us-east-1, eu-west-1, and ap-southeast-1.

bash
# Deploy to cloud (all plans)
aicossic deploy <pipeline-name> \
  --target cloud \
  --region us-east-1 \
  --replicas 2

# Scale existing deployment (Professional+)
aicossic scale <pipeline-name> --replicas 8

On-Premise Deployment Enterprise

Enterprise plans include the aicossic-enterprise-runtime — a self-contained runtime that runs entirely within your network perimeter with no outbound data.

bash
# Pull enterprise runtime (requires Enterprise license key)
docker pull aicossic/enterprise-runtime:2.4.1

# Start with your license
docker run -d \
  -e AICOSSIC_LICENSE="lic_ent_your_key" \
  -e INTFER_ENDPOINT="https://your-intfer-instance" \
  -p 8080:8080 \
  aicossic/enterprise-runtime:2.4.1
🏢
Air-gap support: Enterprise runtime ships with all models pre-bundled and can operate without any internet access. SOC 2 Type II and HIPAA compliance documentation available on request.

Edge Deployment

Deploy lightweight pipeline endpoints to Cloudflare edge nodes via mascom-edge (Professional+). Sub-10ms global latency.

bash
aicossic deploy <pipeline-name> \
  --target edge \
  --provider cloudflare \
  --model-size tiny  # Edge requires models ≤ 100M params

Changelog

v2.4.1 — March 2026

  • Added speculative_decoding option to Intfer.cc backend for 3× LLM throughput
  • Java SDK: builder pattern API, improved async support
  • Node.js SDK: ESM-only, full TypeScript 5.x types
  • Fixed race condition in pipeline autoscale under burst load

v2.4.0 — February 2026

  • Enterprise on-premise runtime GA
  • Intfer.cc int4 quantization support (4× model compression)
  • New models: aicossic/code-13b, aicossic/vision-v2
  • Edge deployment (Cloudflare Workers) support

v2.3.0 — January 2026

  • Python 3.13 support
  • AuthFor SSO integration for enterprise auth
  • Multi-region cloud deployment (eu-west-1, ap-southeast-1)
MASCOM PublicOS WeylandAI GameGob HelmCorp FilmLine HALside News AuthFor