You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
|[OpenAI API](https://openai.com/index/openai-api/)| LLM API for recommender system | Pay-per-use |`OPENAI_API_KEY`|[Quick Start Guide](https://platform.openai.com/docs/quickstart)|
32
+
33
+
# 🎯 Getting Started
34
+
35
+
## 1. Clone the Repository
36
+
37
+
Start by cloning the repository and navigating to the project directory:
To set it up first deactivate any active virtual environment and move to the second-brain-online directory:
4
46
```bash
@@ -40,7 +82,16 @@ After running the doctor command, you should see something like this:
40
82
```
41
83
[More on installing Crawl4AI](https://docs.crawl4ai.com/core/installation/)
42
84
43
-
# 🏗️ Project Structure
85
+
## 3. Environment Configuration
86
+
87
+
Before running any components:
88
+
1. Create your environment file:
89
+
```bash
90
+
cp .env.example .env
91
+
```
92
+
2. Open `.env` and configure the required credentials following the inline comments and the recommendations from the [Cloud Services](#-prerequisites) section.
93
+
94
+
# 📁 Project Structure
44
95
45
96
At Decoding ML we teach how to build production ML systems, thus the course follows the structure of a real-world Python project:
46
97
@@ -63,7 +114,7 @@ At Decoding ML we teach how to build production ML systems, thus the course foll
63
114
└── pyproject.toml # Project dependencies
64
115
```
65
116
66
-
# Infrastructure
117
+
# 🏗️ Set Up Your Local Infrastructure
67
118
68
119
To start the local infrastructure (ZenML, MongoDB):
69
120
```bash
@@ -75,15 +126,15 @@ To stop the local infrastructure (ZenML, MongoDB):
75
126
make local-infrastructure-down
76
127
```
77
128
78
-
# Running the Code / Lessons
129
+
# ⚡️ Running the Code for Each Module
79
130
80
-
## Lesson 1: Build your Second Brain AI assistant
131
+
## Module 1: Build your Second Brain AI assistant
81
132
82
133
Lesson: [Build your Second Brain AI assistant]()
83
134
84
135
No code to run for this lesson. Read the lesson to understand the problem and overall architecture of the Second Brain AI assistant.
85
136
86
-
## Lesson 2: ETL pipeline
137
+
## Module 2: ETL pipeline
87
138
88
139
### Prepare Notion data
89
140
@@ -105,7 +156,7 @@ Run the ETL pipeline to crawl, score and ingest the Notion data into MongoDB:
105
156
make etl-pipeline
106
157
# Validate using test: make test-etl-pipeline
107
158
```
108
-
Running costs: ~$0.5
159
+
Running costs: ~$0.5 </br>
109
160
Running time: ~30 minutes
110
161
111
162
If you want to avoid any costs or waiting times, you can use our pre-computed dataset to populate MongoDB. Also, as crawling can often fail, you can use this dataset to skip the crawling step:
@@ -115,12 +166,12 @@ make download-crawled-dataset
115
166
make etl-precomputed-pipeline
116
167
```
117
168
118
-
## Lesson 3: Generate Fine-tuning Dataset
169
+
## Module 3: Generate Fine-tuning Dataset
119
170
120
171
```bash
121
172
make generate-dataset-pipeline
122
173
```
123
-
Running costs: ~$1.5
174
+
Running costs: ~$1.5 </br>
124
175
Running time: ~60 minutes
125
176
126
177
In case you want to avoid any costs or waiting times, you can use our pre-computed dataset available on Hugging Face, which is already set as default in future steps: [pauliusztin/second_brain_course_summarization_task](https://huggingface.co/datasets/pauliusztin/second_brain_course_summarization_task).
@@ -148,30 +199,31 @@ The Agentic App sits in the online environment, which is implemented as a differ
148
199
149
200
Go to the [apps/second-brain-online](../second-brain-online/) folder and follow the instructions there to set it up and run it.
150
201
151
-
## Utility commands
202
+
#🔧 Utlity Commands
152
203
153
-
###Formatting
204
+
## Formatting
154
205
155
206
```
156
207
make format-check
157
208
make format-fix
158
209
```
159
-
### Linting
210
+
211
+
## Linting
160
212
161
213
```bash
162
214
make lint-check
163
215
make lint-fix
164
216
```
165
217
166
-
###Tests
218
+
## Tests
167
219
168
220
```bash
169
221
make test
170
222
```
171
223
172
-
## Others
224
+
#🧊 Notion (optional)
173
225
174
-
### Notion
226
+
In case you want to use your own Notion data, you can follow these steps to set up an integration and read from your Notion database:
175
227
176
228
1. Go to [https://www.notion.so/profile].
177
229
2. Create an integration following [this tutorial](https://developers.notion.com/docs/authorization).
0 commit comments