使用 Bright Data 抓取即将发生的事件
中级
这是一个AI领域的自动化工作流,包含 11 个节点。主要使用 Code, Html, HttpRequest, GoogleSheets, ScheduleTrigger 等节点,结合人工智能技术实现智能自动化。 使用 Bright Data 和 n8n 的自动化事件发现
前置要求
- •可能需要目标 API 的认证凭证
- •Google Sheets API 凭证
分类
工作流预览
可视化展示节点连接关系,支持缩放和平移
导出工作流
复制以下 JSON 配置到 n8n 导入,即可使用此工作流
{
"id": "NY5tq9f8iYwpvPC6",
"meta": {
"instanceId": "60046904b104f0f72b2629a9d88fe9f676be4035769f1f08dad1dd38a76b9480"
},
"name": "Scrape_Upcoming_Events_using_bright_data",
"tags": [],
"nodes": [
{
"id": "5012cf3e-7fa6-4971-906b-760baeb51396",
"name": "Trigger - Weekly Run",
"type": "n8n-nodes-base.scheduleTrigger",
"position": [
-440,
1160
],
"parameters": {
"rule": {
"interval": [
{
"field": "weeks",
"triggerAtDay": [
1
],
"triggerAtHour": 8
}
]
}
},
"typeVersion": 1.2
},
{
"id": "8962483b-d095-4ade-bf9c-c3bfa5fe1831",
"name": "Scrape event website using bright data",
"type": "n8n-nodes-base.httpRequest",
"position": [
-200,
1160
],
"parameters": {
"url": "https://api.brightdata.com/request",
"method": "POST",
"options": {},
"sendBody": true,
"sendHeaders": true,
"bodyParameters": {
"parameters": [
{
"name": "zone",
"value": "n8n_unblocker"
},
{
"name": "url",
"value": "https://www.eventbrite.com/d/online/technology--events/"
},
{
"name": "country",
"value": "us"
},
{
"name": "format",
"value": "raw"
}
]
},
"headerParameters": {
"parameters": [
{
"name": "Authorization",
"value": "Bearer API_KEY"
}
]
}
},
"typeVersion": 4.2
},
{
"id": "12d62c8b-047d-4890-85f7-c9a8097fcc2d",
"name": "Parse HTML - Extract Event Cards",
"type": "n8n-nodes-base.html",
"position": [
60,
1160
],
"parameters": {
"options": {},
"operation": "extractHtmlContent",
"extractionValues": {
"values": [
{
"key": "Title",
"cssSelector": "h3",
"returnArray": true
},
{
"key": "Date and Time",
"cssSelector": "div.Stack_root__1ksk7 > p:nth-of-type(1)",
"returnArray": true
}
]
}
},
"typeVersion": 1.2
},
{
"id": "df0d0e23-1afe-4c74-b408-e5b3123c13a4",
"name": "Format Event Data",
"type": "n8n-nodes-base.code",
"position": [
280,
1160
],
"parameters": {
"jsCode": "// Get the input data from the previous node\nconst inputData = items[0].json;\n\n// Access the arrays from the input object\nconst titles = inputData.Title;\nconst dates = inputData['Date and Time'];\n\n// Use a Set to track titles we've already processed to remove duplicates\nconst seenTitles = new Set();\nconst cleanedEvents = [];\n\n// Get the number of actual events (the input has duplicates)\nconst eventCount = dates.length;\n\nfor (let i = 0; i < eventCount; i++) {\n const title = titles[i];\n const date = dates[i];\n\n // 1. Skip if the title is empty or we've already processed this event\n if (!title || seenTitles.has(title)) {\n continue;\n }\n\n // 2. Filter out irrelevant \"Trends\" items that start with a number (e.g., \"1. Tickets\")\n if (/^\\d+\\.\\s/.test(title)) {\n continue;\n }\n \n // 3. Add the unique event to our results (without the URL)\n cleanedEvents.push({\n \"Title\": title,\n \"Date and Time\": date\n });\n\n // 4. Mark this title as seen to avoid adding it again\n seenTitles.add(title);\n}\n\n// Return the newly structured and cleaned array of event objects\nreturn cleanedEvents;"
},
"typeVersion": 2
},
{
"id": "c58d917f-d756-45ad-a38a-b52254250154",
"name": "Save to Google Sheet",
"type": "n8n-nodes-base.googleSheets",
"position": [
580,
1160
],
"parameters": {
"columns": {
"value": {
"Title": "={{ $json.Title }}",
"Date & Time": "={{ $json['Date and Time'] }}"
},
"schema": [
{
"id": "Title",
"type": "string",
"display": true,
"required": false,
"displayName": "Title",
"defaultMatch": false,
"canBeUsedToMatch": true
},
{
"id": "Date & Time",
"type": "string",
"display": true,
"required": false,
"displayName": "Date & Time",
"defaultMatch": false,
"canBeUsedToMatch": true
}
],
"mappingMode": "defineBelow",
"matchingColumns": [],
"attemptToConvertTypes": false,
"convertFieldsToString": false
},
"options": {},
"operation": "append",
"sheetName": {
"__rl": true,
"mode": "list",
"value": "gid=0",
"cachedResultUrl": "https://docs.google.com/spreadsheets/d/157HRnzYP9IShr4jTQH7_y3r35cq2NVu0hv7kAW9kqn0/edit#gid=0",
"cachedResultName": "Sheet1"
},
"documentId": {
"__rl": true,
"mode": "list",
"value": "157HRnzYP9IShr4jTQH7_y3r35cq2NVu0hv7kAW9kqn0",
"cachedResultUrl": "https://docs.google.com/spreadsheets/d/157HRnzYP9IShr4jTQH7_y3r35cq2NVu0hv7kAW9kqn0/edit?usp=drivesdk",
"cachedResultName": "Events"
}
},
"credentials": {
"googleSheetsOAuth2Api": {
"id": "r2mDaisH6e9VkwHl",
"name": "Google Sheets account"
}
},
"typeVersion": 4.6
},
{
"id": "ed7d1601-99b4-4191-b75c-cb7060a42bc5",
"name": "Sticky Note",
"type": "n8n-nodes-base.stickyNote",
"position": [
-480,
460
],
"parameters": {
"color": 5,
"width": 440,
"height": 880,
"content": "## 🧩 SECTION 1: 🔄 **Trigger + Scrape Website Content**\n\n### 🧭 Purpose: Kickstart the automation & fetch data from the web\n\n---\n\n| 🔧 **Nodes Involved** | 🔹 `Schedule Trigger`<br>🔹 `HTTP Request (Bright Data Web Unlocker)` |\n| --------------------- | ----------------------------------------------------------------------- |\n| 📌 **Goal** | Automatically fetch webinar details from Eventbrite (or any event site) |\n| 🛠️ **How it works** | |\n\n1. **⏰ `Schedule Trigger`**\n This node runs the entire workflow at a preset time (e.g., daily at 8 AM). It ensures your automation runs hands-free — no manual clicking.\n\n2. **🌐 `HTTP Request`**\n This node uses **Bright Data's Web Unlocker** to bypass bot detection and scrape real event pages from protected websites like Eventbrite.\n The **POST** request is sent to `https://api.brightdata.com/request`, which returns **raw HTML** of the event listings page.\n\n💡 *Why Bright Data?*\nWebsites like Eventbrite use anti-bot systems. Bright Data safely navigates that with proxies + human-like browsing.\n\n---\n\n"
},
"typeVersion": 1
},
{
"id": "24726d95-f802-45e8-868d-575ce6453580",
"name": "Sticky Note1",
"type": "n8n-nodes-base.stickyNote",
"position": [
0,
0
],
"parameters": {
"color": 6,
"width": 420,
"height": 1340,
"content": "## 🧩 SECTION 2: 🔍 **Extract & Structure Event Data**\n\n### 🧠 Purpose: Turn messy HTML into clean, usable event info\n\n---\n\n| 🔧 **Nodes Involved** | 🔹 `HTML Extract`<br>🔹 `Code Node` |\n| --------------------- | -------------------------------------------------------- |\n| 📌 **Goal** | Isolate event titles, times, and links from the raw HTML |\n| 🛠️ **How it works** | |\n\n1. **🧾 `HTML Extract`**\n This node lets you select elements from the HTML using **CSS selectors**, just like a web developer would.\n You extract:\n\n * `.eds-event-card-content__title` → **Title**\n * `.eds-event-card-content__sub-title` → **Date & Time**\n * `.eds-event-card-content__action-link[href]` → **Event URL**\n\n2. **🧮 `Code Node`**\n Here, we **loop through the extracted data** and format it into clean JSON objects:\n\n ```js\n return items[0].json.titles.map((title, i) => {\n return {\n json: {\n title,\n date: items[0].json.dates[i],\n link: items[0].json.links[i]\n }\n };\n });\n ```\n\n Result: You now have a list of clean, structured webinar entries.\n\n💡 *Why use a code node?*\nIt transforms raw scraped chunks into usable pieces ready for storage or integration.\n\n---\n\n"
},
"typeVersion": 1
},
{
"id": "834b6789-b284-49e0-86ba-a9d803fb52aa",
"name": "Sticky Note2",
"type": "n8n-nodes-base.stickyNote",
"position": [
460,
540
],
"parameters": {
"color": 3,
"width": 340,
"height": 800,
"content": "## 🧩 SECTION 3: 📄 **Store Events in Google Sheets**\n\n### 💾 Purpose: Save your scraped webinars into a spreadsheet\n\n---\n\n| 🔧 **Node Involved** | 🔹 `Google Sheets: Append` |\n| -------------------- | ----------------------------------------------- |\n| 📌 **Goal** | Automatically log each event into a spreadsheet |\n| 🛠️ **How it works** | |\n\n* This node appends each structured event entry to a connected **Google Sheet**.\n* Each row includes:\n 📌 `Event Title` | 📅 `Date & Time` | 🔗 `URL`\n* You can filter, sort, or share this sheet as a **database of upcoming webinars**.\n\n💡 *Why Google Sheets?*\nIt’s universal, easy to search/filter, and sharable with a team. No special tool or database needed.\n\n---\n\n"
},
"typeVersion": 1
},
{
"id": "48ecec93-c46c-4398-ab14-bcfe617b0728",
"name": "Sticky Note9",
"type": "n8n-nodes-base.stickyNote",
"position": [
-2140,
480
],
"parameters": {
"color": 4,
"width": 1300,
"height": 320,
"content": "=======================================\n WORKFLOW ASSISTANCE\n=======================================\nFor any questions or support, please contact:\n Yaron@nofluff.online\n\nExplore more tips and tutorials here:\n - YouTube: https://www.youtube.com/@YaronBeen/videos\n - LinkedIn: https://www.linkedin.com/in/yaronbeen/\n=======================================\n"
},
"typeVersion": 1
},
{
"id": "4bceecb8-56d2-47d7-8833-92f29af8f2fa",
"name": "Sticky Note3",
"type": "n8n-nodes-base.stickyNote",
"position": [
-2140,
820
],
"parameters": {
"color": 4,
"width": 1289,
"height": 2298,
"content": "### 🎓 **Auto-Webinar Harvester**\n\n#### *Scrape Upcoming Events & Save to Google Sheets Automatically*\n\n> **Tagline:** “Your always-on assistant for discovering industry webinars — no clicks required.”\n> 🧠💻⚙️📅\n\n---\n\n## 🧩 SECTION 1: 🔄 **Trigger + Scrape Website Content**\n\n### 🧭 Purpose: Kickstart the automation & fetch data from the web\n\n---\n\n| 🔧 **Nodes Involved** | 🔹 `Schedule Trigger`<br>🔹 `HTTP Request (Bright Data Web Unlocker)` |\n| --------------------- | ----------------------------------------------------------------------- |\n| 📌 **Goal** | Automatically fetch webinar details from Eventbrite (or any event site) |\n| 🛠️ **How it works** | |\n\n1. **⏰ `Schedule Trigger`**\n This node runs the entire workflow at a preset time (e.g., daily at 8 AM). It ensures your automation runs hands-free — no manual clicking.\n\n2. **🌐 `HTTP Request`**\n This node uses **Bright Data's Web Unlocker** to bypass bot detection and scrape real event pages from protected websites like Eventbrite.\n The **POST** request is sent to `https://api.brightdata.com/request`, which returns **raw HTML** of the event listings page.\n\n💡 *Why Bright Data?*\nWebsites like Eventbrite use anti-bot systems. Bright Data safely navigates that with proxies + human-like browsing.\n\n---\n\n## 🧩 SECTION 2: 🔍 **Extract & Structure Event Data**\n\n### 🧠 Purpose: Turn messy HTML into clean, usable event info\n\n---\n\n| 🔧 **Nodes Involved** | 🔹 `HTML Extract`<br>🔹 `Code Node` |\n| --------------------- | -------------------------------------------------------- |\n| 📌 **Goal** | Isolate event titles, times, and links from the raw HTML |\n| 🛠️ **How it works** | |\n\n1. **🧾 `HTML Extract`**\n This node lets you select elements from the HTML using **CSS selectors**, just like a web developer would.\n You extract:\n\n * `.eds-event-card-content__title` → **Title**\n * `.eds-event-card-content__sub-title` → **Date & Time**\n * `.eds-event-card-content__action-link[href]` → **Event URL**\n\n2. **🧮 `Code Node`**\n Here, we **loop through the extracted data** and format it into clean JSON objects:\n\n ```js\n return items[0].json.titles.map((title, i) => {\n return {\n json: {\n title,\n date: items[0].json.dates[i],\n link: items[0].json.links[i]\n }\n };\n });\n ```\n\n Result: You now have a list of clean, structured webinar entries.\n\n💡 *Why use a code node?*\nIt transforms raw scraped chunks into usable pieces ready for storage or integration.\n\n---\n\n## 🧩 SECTION 3: 📄 **Store Events in Google Sheets**\n\n### 💾 Purpose: Save your scraped webinars into a spreadsheet\n\n---\n\n| 🔧 **Node Involved** | 🔹 `Google Sheets: Append` |\n| -------------------- | ----------------------------------------------- |\n| 📌 **Goal** | Automatically log each event into a spreadsheet |\n| 🛠️ **How it works** | |\n\n* This node appends each structured event entry to a connected **Google Sheet**.\n* Each row includes:\n 📌 `Event Title` | 📅 `Date & Time` | 🔗 `URL`\n* You can filter, sort, or share this sheet as a **database of upcoming webinars**.\n\n💡 *Why Google Sheets?*\nIt’s universal, easy to search/filter, and sharable with a team. No special tool or database needed.\n\n---\n\n## 🎯 Why This Workflow Is Powerful (Even for Beginners)\n\n| 🚀 Benefit | ✅ How It Helps You |\n| --------------------------------- | ----------------------------------------------------------------------- |\n| Zero manual scraping | Just set a schedule and forget — everything updates on its own |\n| Works with hard-to-scrape sites | Bright Data + Web Unlocker beats CAPTCHAs and blocks |\n| Clean, structured output | Easy to use data for marketing, lead gen, or syncing to Google Calendar |\n| Google Sheet as a source of truth | Everyone on your team can access and use the data |\n\n---\n\n"
},
"typeVersion": 1
},
{
"id": "197d2ec1-609e-4905-8671-e41ed6ceb92f",
"name": "Sticky Note4",
"type": "n8n-nodes-base.stickyNote",
"position": [
900,
540
],
"parameters": {
"color": 7,
"width": 380,
"height": 240,
"content": "## I’ll receive a tiny commission if you join Bright Data through this link—thanks for fueling more free content!\n\n### https://get.brightdata.com/1tndi4600b25"
},
"typeVersion": 1
}
],
"active": false,
"pinData": {},
"settings": {
"executionOrder": "v1"
},
"versionId": "b59d50da-e665-47a2-975b-432971342df5",
"connections": {
"df0d0e23-1afe-4c74-b408-e5b3123c13a4": {
"main": [
[
{
"node": "c58d917f-d756-45ad-a38a-b52254250154",
"type": "main",
"index": 0
}
]
]
},
"5012cf3e-7fa6-4971-906b-760baeb51396": {
"main": [
[
{
"node": "8962483b-d095-4ade-bf9c-c3bfa5fe1831",
"type": "main",
"index": 0
}
]
]
},
"12d62c8b-047d-4890-85f7-c9a8097fcc2d": {
"main": [
[
{
"node": "df0d0e23-1afe-4c74-b408-e5b3123c13a4",
"type": "main",
"index": 0
}
]
]
},
"8962483b-d095-4ade-bf9c-c3bfa5fe1831": {
"main": [
[
{
"node": "12d62c8b-047d-4890-85f7-c9a8097fcc2d",
"type": "main",
"index": 0
}
]
]
}
}
}常见问题
如何使用这个工作流?
复制上方的 JSON 配置代码,在您的 n8n 实例中创建新工作流并选择「从 JSON 导入」,粘贴配置后根据需要修改凭证设置即可。
这个工作流适合什么场景?
中级 - 人工智能
需要付费吗?
本工作流完全免费,您可以直接导入使用。但请注意,工作流中使用的第三方服务(如 OpenAI API)可能需要您自行付费。
相关工作流推荐
通过Bright Data进行竞争对手价格监控
使用Bright Data和n8n的自动竞争对手价格监控
If
Code
Html
+
If
Code
Html
15 节点Yaron Been
人工智能
自动化论坛监控_via_Bright_data
使用Bright Data和n8n的论坛监控自动化
Set
Code
Html
+
Set
Code
Html
17 节点Yaron Been
人工智能
研究论文爬虫到Google Sheets
使用Bright Data和n8n自动化研究论文收集
Set
Code
Html
+
Set
Code
Html
12 节点Yaron Been
人工智能
CrunchBase 投资者数据
自动化投资者情报:CrunchBase到Google Sheets数据采集器
Code
Http Request
Google Sheets
+
Code
Http Request
Google Sheets
8 节点Yaron Been
财务
自动同步本地事件到Google日历
使用n8n自动同步本地事件到Google日历
Code
Html
Http Request
+
Code
Html
Http Request
11 节点Yaron Been
人工智能
通过 Bright Data 的自动化自由职业零工抓取器
使用Bright Data和n8n的自动化自由职业零工查找器
Set
Html
Http Request
+
Set
Html
Http Request
11 节点Yaron Been
人工智能
工作流信息
难度等级
中级
节点数量11
分类1
节点类型6
作者
Yaron Been
@yaron-nofluffBuilding AI Agents and Automations | Growth Marketer | Entrepreneur | Book Author & Podcast Host If you need any help with Automations, feel free to reach out via linkedin: https://www.linkedin.com/in/yaronbeen/ And check out my Youtube channel: https://www.youtube.com/@YaronBeen/videos
外部链接
在 n8n.io 查看 →
分享此工作流