3.3 KiB
slug, title, type, description, shortDescription, publishedAt, readingTime, status, tags, icon
| slug | title | type | description | shortDescription | publishedAt | readingTime | status | tags | icon | ||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| n8n-automations | n8n Automations | Academic Project | An academic project exploring the automation of GenAI workflows using n8n and Ollama for self-hosted AI applications, including personalized research agents and productivity hubs. | Automating GenAI workflows with n8n and Ollama in a self-hosted environment. | 2026-03-15 | 2 | In progress |
|
i-ph-plugs-connected-duotone |
Overview
This project focuses on designing and implementing autonomous workflows that leverage Large Language Models (LLMs) to streamline productivity and academic research. By orchestrating Generative AI through a self-hosted infrastructure on my ArtLab, I built a private ecosystem that acts as both a personal assistant and a specialized research agent.
Key Workflows
1. Centralized Productivity Hub
I developed a synchronization engine that bridges Notion, Google Calendar, and Todoist.
- Contextual Sync: Academic events, such as course schedules and exam dates, are pulled from Notion and reflected in my calendar and task manager.
- Daily Briefing: Every morning, the system triggers a workflow that compiles my schedule, pending tasks, and a local weather report into a single, centralized email summary. This ensures a frictionless start to the day with all critical information in one place.
2. Intelligent Research Engine (RSS & RAG)
To stay at the forefront of AI research, I built an automated pipeline for academic and technical monitoring.
- Multi-Source Fetching: The system monitors RSS feeds from arXiv, Hugging Face, Hacker News, selfho.st, and major industry blogs (OpenAI, Google Research, Meta).
- Semantic Filtering: Using LLMs, articles are filtered and ranked based on my specific research profile, with a focus on robust distributed learning.
- Knowledge Base: Relevant papers and posts are automatically stored in a structured Notion database.
- Interactive Research Agent: I integrated a chat interface within n8n that allows me to query this collected data. I can request summaries, ask specific technical questions about a paper, or extract the most relevant insights for my current thesis work.
Technical Architecture
The environment is built to handle complex multi-step chains, moving beyond simple API calls to create context-aware agents.
Integrated Ecosystem
- Intelligence Layer: Integration with Gemini (API) and Ollama (local) for summarization and semantic sorting.
- Data Sources: RSS feeds and Notion databases.
- Notifications & UI: Gmail for briefings and Discord for real-time system alerts.
Key Objectives
- Privacy-Centric AI: Ensuring that sensitive academic data and personal schedules remain within a self-hosted or controlled environment.
- Academic Efficiency: Reducing the "noise" of information overload by using AI to surface only the most relevant research papers.
- Low-Code Orchestration: Utilizing n8n to manage complex logic and API interactions without the overhead of maintaining a massive custom codebase.
This project is currently under active development as I refine the RAG (Retrieval-Augmented Generation) logic and optimize the filtering prompts for my research.