The .txt Python library provides access to the .txt REST API from Python 3.10+ applications.
It provides two client surfaces:
DotTxtfor sync access with DotTxt helpers and OpenAI-compatible namespacesAsyncDotTxtfor async access with the same helper semantics
The API uses:
- base URL:
https://api.dottxt.ai/v1 - auth:
Authorization: Bearer $DOTTXT_API_KEY - primary endpoints:
GET /modelsandPOST /chat/completions
pip install dottxtexport DOTTXT_API_KEY="your-api-key"All clients in this package read DOTTXT_API_KEY by default.
Optional overrides:
export DOTTXT_BASE_URL="https://api.dottxt.ai/v1"
export DOTTXT_MODEL="<model-id>"Use the dottxt CLI for login, model discovery, and one-off generation.
- CLI reference: docs/cli.md
- Client reference: docs/client.md
Choose the client that matches the shape you want to work with:
DotTxt.generate(...)andAsyncDotTxt.generate(...)accept JSON Schema as a string/object, plus any typed schema supported by PydanticTypeAdapter(for example: Pydantic models, Enums, Literals, Unions, Optionals, and typed containers), and objects exposingto_json() -> str(for example Genson). Root{"type":"structural-tag", ...}schema objects are accepted without JSON Schema metaschema validation. They return a validated Pydantic model instance for Pydantic input, or parsed JSON for the other schema input types.DotTxtandAsyncDotTxtalso expose OpenAI SDKchatandmodelsnamespaces for direct SDK access alongside DotTxt helpers.
For constructor kwargs passed through to the OpenAI SDK client
(DotTxt(..., **client_kwargs) / AsyncDotTxt(..., **client_kwargs)), see
OpenAI Python base client parameters.
from typing import Literal
from pydantic import BaseModel, Field
from dottxt import DotTxt
class IncidentSummary(BaseModel):
severity: Literal["low", "medium", "high"]
team: str = Field(max_length=32)
client = DotTxt()
result = client.generate(
model="openai/gpt-oss-20b",
input="Summarize this incident: checkout errors are blocking purchases.",
response_format=IncidentSummary,
)
print(result)
# Example model output:
# severity='high' team='checkout'
print(result.model_dump())
# Example output:
# {'severity': 'high', 'team': 'checkout'}
models = client.models.list()
print([model.id for model in models.data])
# Example output:
# ['openai/gpt-oss-20b', 'openai/gpt-4.1-mini']import asyncio
from typing import Literal
from pydantic import BaseModel, Field
from dottxt import AsyncDotTxt
class IncidentSummary(BaseModel):
severity: Literal["low", "medium", "high"]
team: str = Field(max_length=32)
async def main() -> None:
client = AsyncDotTxt()
result = await client.generate(
model="openai/gpt-oss-20b",
input="Summarize this incident: checkout errors are blocking purchases.",
response_format=IncidentSummary,
)
print(result)
# Example model output:
# severity='high' team='checkout'
print(result.model_dump())
# Example output:
# {'severity': 'high', 'team': 'checkout'}
models = await client.models.list()
print([model.id for model in models.data])
# Example output:
# ['openai/gpt-oss-20b', 'openai/gpt-4.1-mini']
asyncio.run(main())For DotTxt and AsyncDotTxt, generate(...) accepts response_format as:
- a Pydantic model class
- a TypedDict type
- a dataclass type
- an Enum class
- a
typing.Literal[...]type - a
typing.Union[...]type - a
typing.Optional[...]type - typed containers such as
list[...],dict[...],tuple[...] - a JSON string containing JSON Schema
- a JSON object (
dict) - an object exposing
to_json() -> strthat returns JSON Schema
Notes:
- Raw list instances as
response_formatare not supported. - Root
{"type":"structural-tag", ...}schema objects bypass metaschema checks.
For direct chat.completions.create(...), pass the wrapped OpenAI-style
response_format payload yourself.
Use DotTxt.models.list() and AsyncDotTxt.models.list() for model listing.
Use DotTxt when you want an OpenAI-style client surface with
chat.completions.create(...) and models.list().
from typing import Literal
from pydantic import BaseModel, Field
from dottxt import DotTxt as OpenAI
class IncidentSummary(BaseModel):
severity: Literal["low", "medium", "high"]
team: str = Field(max_length=32)
client = OpenAI()
completion = client.chat.completions.create(
model="openai/gpt-oss-20b",
messages=[
{
"role": "user",
"content": "Summarize this incident: checkout errors are blocking purchases.",
}
],
response_format={
"type": "json_schema",
"json_schema": {
"name": "incident_summary",
"schema": IncidentSummary.model_json_schema(),
},
},
)
print(completion.choices[0].message.content)
# Example output:
# {"severity":"high","team":"checkout"}
models = client.models.list()
print([model.id for model in models.data])
# Example output:
# ['openai/gpt-oss-20b', 'openai/gpt-4.1-mini']The compatibility surface expects the wrapped OpenAI-style
response_format payload:
{"type": "json_schema", "json_schema": {...}}