Hey there! I'm AIt (acts like "alt" /ɔːlt/, but also pronounced as "eight" /eɪt/). It depends. 🤷♂️
Thinking... 🤔
-
Connectors:
- GitHub integration for repository analysis and OAuth 2.0 authentication
- Spotify integration for music insights and playlist analysis
- X integration for tweets analysis
- Modular connector architecture with shared utilities
- Type-safe OpenAPI generated interfaces
- Automatic token refresh and persistence
-
ETL Pipeline:
- Extract data from multiple sources using typed connectors
- Transform using LangChain and multiple embedding options
- Load into vector databases for semantic search
- Support for both Python and Node.js embedding generation
- Flexible pipeline configuration
-
Storage Solutions:
- PostgreSQL for structured data and OAuth tokens
- Qdrant for vector similarity search
- Ollama for local LLM processing (deepseek-r1:1.5b)
- Redis for job queue and caching
-
Scheduler:
- Schedule and manage ETL tasks with BullMQ
- Automated token refresh and data synchronization
- Supports cron expressions for periodic tasks
- Configurable job priorities and retries
- Install Node.js dependencies:
corepack enable
pnpm install
- Start required services:
It requires Docker and Docker Compose to be installed.
pnpm start:services # Starts PostgreSQL, Qdrant, Ollama, etc
- Set up environment variables:
You can follow the
.env.example
file to create your own.env
file. The project also supports.env.test
for testing purposes.
# Database Configuration
POSTGRES_URL=postgresql://root:toor@localhost:5432/ait
# GitHub OAuth
GITHUB_CLIENT_ID=your_github_client_id
GITHUB_CLIENT_SECRET=your_github_secret
# Spotify OAuth
SPOTIFY_CLIENT_ID=your_spotify_client_id
SPOTIFY_CLIENT_SECRET=your_spotify_secret
# X OAuth
X_CLIENT_ID=your_x_client_id
X_CLIENT_SECRET=your_x_secret
- Initialize the database:
# Ensure you have set the required environment variables
pnpm migrate
AIt provides flexibility in running the ETL process either automatically through the Scheduler or manually as needed.
The Scheduler manages also the ETL process. It uses BullMQ for job queue management and supports cron expressions for periodic tasks.
docker compose build ait_scheduler && docker compose up -d ait_scheduler
Ensure the Scheduler is properly configured by setting the necessary environment variables in .env
, and the following services are running:
- ait_postgres
- ait_qdrant
- ait_redis
If you prefer to run the ETL process manually, you can do so by following these steps:
cd packages/transformers/retove
# Ensure you have set the required environment variables
pnpm etl
After the ETL process, you can generate text using the TextGeneration
service. Here's how to get started:
cd packages/infrastructure/langchain # <- The service will move out of this folder soon
npx tsx src/services/text-generation/text-generation.service.e2e.ts
The E2E tests will:
- Connect to your Qdrant collections (github_repositories_collection and spotify_tracks_collection)
- Generate embeddings for test prompts
- Perform similarity searches
- Generate responses using Ollama (deepseek-r1:1.5b)
The project provides smart connectors for GitHub, Spotify, X, and more through a unified gateway. Here's how to get started:
First, generate the TypeScript interfaces from OpenAPI specifications:
cd packages/connectors
pnpm generate:openapi
Note: Generated types are not committed to avoid repository bloat.
cd packages/gateway
pnpm dev
AIt securely connects to platforms using OAuth 2.0. Visit these URLs to authenticate:
- GitHub:
http://localhost:3000/api/github/auth
- Spotify:
http://localhost:3000/api/spotify/auth
- X:
http://localhost:3000/api/x/auth
Once authenticated, AIt securely stores and manages OAuth tokens in the database for future requests.
AIt uses Ollama for local LLM processing. Here's how to set it up:
- Start the Ollama service:
docker compose up -d ait_ollama
- Install the model:
docker exec -it ait_ollama sh -c "ollama pull deepseek-r1:1.5b"
The model is used for:
- Generating embeddings via LangChain
- Text generation and analysis
- Semantic search operations
Run tests in an isolated environment using Docker Compose:
# Run tests (services, migrations and seeding happen automatically)
pnpm test
Note: Ensure the
ait_testing
database is properly initialized. The project uses.env.test
for test configuration.
# Generate OpenAPI types
pnpm generate:openapi
# Generate database types
cd packages/infrastructure/postgres
pnpm db:generate
# Run migrations
pnpm db:migrate
# Access database UI
pnpm db:studio
pnpm lint # Run linting
pnpm lint:fix # Fix linting issues