ScrapegraphAIを使ってニュース記事を自動スクラップしてGoogleシートに保存
中級
これはMarket Research, AI Summarization分野の自動化ワークフローで、8個のノードを含みます。主にCode, GoogleSheets, ScheduleTrigger, ScrapegraphAiなどのノードを使用。 ScrapegraphAIを使ってニュース記事を自動スクレイピングし、Google Sheetsに保存
前提条件
- •Google Sheets API認証情報
ワークフロープレビュー
ノード接続関係を可視化、ズームとパンをサポート
ワークフローをエクスポート
以下のJSON設定をn8nにインポートして、このワークフローを使用できます
{
"id": "MIllJmbqayQrZM1F",
"meta": {
"instanceId": "521567c5f495f323b77849c4cfd0c9f4f2396c986e324e0e66c8425b6f124744",
"templateCredsSetupCompleted": true
},
"name": "Automate News Article Scraping with ScrapegraphAI and Store in Google Sheets",
"tags": [],
"nodes": [
{
"id": "37df323b-5c75-495f-ba19-b8642c02d96f",
"name": "Automated News Collection Trigger",
"type": "n8n-nodes-base.scheduleTrigger",
"position": [
700,
820
],
"parameters": {
"rule": {
"interval": [
{}
]
}
},
"typeVersion": 1.2
},
{
"id": "efd61ca5-e248-4027-b705-6d9c5dabe820",
"name": "AI-Powered News Article Scraper",
"type": "n8n-nodes-scrapegraphai.scrapegraphAi",
"position": [
1380,
820
],
"parameters": {
"userPrompt": "Extract all the articles from this site. Use the following schema for response { \"request_id\": \"5a9de102-8a43-4e89-8aae-397c9ca80a9b\", \"status\": \"completed\", \"website_url\": \"https://www.bbc.com/\", \"user_prompt\": \"Extract all the articles from this site.\", \"title\": \"'My friend died right in front of me' - Student describes moment air force jet crashed into school\", \"url\": \"https://www.bbc.com/news/articles/cglzw8y5wy5o\", \"category\": \"Asia\" }",
"websiteUrl": "https://www.bbc.com/"
},
"credentials": {
"scrapegraphAIApi": {
"id": "",
"name": ""
}
},
"typeVersion": 1
},
{
"id": "976d9123-7585-4700-9972-5b2838571a44",
"name": "Google スプレッドシート News Storage",
"type": "n8n-nodes-base.googleSheets",
"position": [
2980,
820
],
"parameters": {
"columns": {
"value": {},
"schema": [
{
"id": "title",
"type": "string",
"display": true,
"removed": false,
"required": false,
"displayName": "title",
"defaultMatch": false,
"canBeUsedToMatch": true
},
{
"id": "url",
"type": "string",
"display": true,
"removed": false,
"required": false,
"displayName": "url",
"defaultMatch": false,
"canBeUsedToMatch": true
},
{
"id": "category",
"type": "string",
"display": true,
"removed": false,
"required": false,
"displayName": "category",
"defaultMatch": false,
"canBeUsedToMatch": true
}
],
"mappingMode": "autoMapInputData",
"matchingColumns": []
},
"options": {},
"operation": "append",
"sheetName": {
"__rl": true,
"mode": "name",
"value": "Sheet1"
},
"documentId": {
"__rl": true,
"mode": "url",
"value": ""
}
},
"credentials": {
"googleSheetsOAuth2Api": {
"id": "",
"name": ""
}
},
"typeVersion": 4.5
},
{
"id": "6d11ae64-e2f8-47ed-854a-c749881ce72c",
"name": "News Data Formatting and Processing",
"type": "n8n-nodes-base.code",
"notes": "Hey this is where \nyou \nformat results ",
"position": [
2140,
820
],
"parameters": {
"jsCode": "// Get the input data\nconst inputData = $input.all()[0].json;\n\n// Extract articles array\nconst articles = inputData.result.articles;\n\n// Map each article and return only title, url, category\nreturn articles.map(article => ({\n json: {\n title: article.title,\n url: article.url,\n category: article.category\n }\n}));"
},
"notesInFlow": true,
"typeVersion": 2
},
{
"id": "ca78baaf-0480-490d-aa9a-3663ca93f5d0",
"name": "付箋1",
"type": "n8n-nodes-base.stickyNote",
"position": [
1180,
460
],
"parameters": {
"color": 5,
"width": 574.9363634768473,
"height": 530.4701664623029,
"content": "# Step 2: AI-Powered News Article Scraper 🤖\n\nThis is the core node which uses ScrapeGraphAI to intelligently extract news articles from any website.\n\n## How to Use\n- Configure the target news website URL\n- Use natural language to describe what data to extract\n- The AI will automatically parse and structure the results\n\n## Configuration\n- **Website URL**: Target news website (e.g., BBC, CNN, Reuters)\n- **User Prompt**: Natural language instructions for data extraction\n- **API Credentials**: ScrapeGraphAI API key required\n\n## Example\n- **Website**: BBC News homepage\n- **Instruction**: \"Extract all article titles, URLs, and categories\"\n\n⚠️ **Note**: This is a community node requiring self-hosting"
},
"typeVersion": 1
},
{
"id": "51a1337b-6a50-43a5-8d6f-8345bc771c7b",
"name": "付箋2",
"type": "n8n-nodes-base.stickyNote",
"position": [
1920,
460
],
"parameters": {
"color": 5,
"width": 574.9363634768473,
"height": 530.4701664623029,
"content": "# Step 3: News Data Formatting and Processing 🧱\n\nThis node transforms and structures the scraped news data for optimal Google Sheets compatibility.\n\n## What it does\n- Extracts articles array from ScrapeGraphAI response\n- Maps each article to standardized format\n- Ensures data consistency and structure\n- Prepares clean data for spreadsheet storage\n\n## Data Structure\n- **title**: Article headline and title\n- **url**: Direct link to the article\n- **category**: Article category or section\n\n## Customization\n- Modify the JavaScript code to extract additional fields\n- Add data validation and cleaning logic\n- Implement error handling for malformed data"
},
"typeVersion": 1
},
{
"id": "2e8cde8e-f534-4f37-a1f9-bcf0fe0b09f9",
"name": "付箋3",
"type": "n8n-nodes-base.stickyNote",
"position": [
460,
460
],
"parameters": {
"color": 5,
"width": 574.9363634768473,
"height": 530.4701664623029,
"content": "# Step 1: Automated News Collection Trigger ⏱️\n\nThis trigger automatically invokes the workflow at specified intervals to collect fresh news content.\n\n## Configuration Options\n- **Frequency**: Daily, hourly, or custom intervals\n- **Time Zone**: Configure for your business hours\n- **Execution Time**: Choose optimal times for news collection\n\n## Best Practices\n- Set appropriate intervals to respect rate limits\n- Consider news website update frequencies\n- Monitor execution logs for any issues\n- Adjust frequency based on your monitoring needs"
},
"typeVersion": 1
},
{
"id": "5606537c-a531-490a-b4ff-6d0dc5e642b4",
"name": "付箋",
"type": "n8n-nodes-base.stickyNote",
"position": [
2680,
460
],
"parameters": {
"color": 5,
"width": 574.9363634768473,
"height": 530.4701664623029,
"content": "# Step 4: Google Sheets News Storage 📊\n\nThis node securely stores the processed news article data in your Google Sheets for analysis and tracking.\n\n## What it does\n- Connects to your Google Sheets account via OAuth2\n- Appends new article data as rows\n- Maintains historical data for trend analysis\n- Provides structured data for business intelligence\n\n## Configuration\n- **Spreadsheet**: Select or create target Google Sheets document\n- **Sheet Name**: Configure worksheet (default: Sheet1)\n- **Operation**: Append mode for continuous data collection\n- **Column Mapping**: Automatic mapping of title, url, category fields\n\n## Data Management\n- Each execution adds new article entries\n- Historical data preserved for analysis\n- Easy export and sharing capabilities\n- Built-in Google Sheets analytics and filtering"
},
"typeVersion": 1
}
],
"active": false,
"pinData": {},
"settings": {
"executionOrder": "v1"
},
"versionId": "c2fee060-f99e-48aa-a280-ac5492715fd9",
"connections": {
"efd61ca5-e248-4027-b705-6d9c5dabe820": {
"main": [
[
{
"node": "6d11ae64-e2f8-47ed-854a-c749881ce72c",
"type": "main",
"index": 0
}
]
]
},
"37df323b-5c75-495f-ba19-b8642c02d96f": {
"main": [
[
{
"node": "efd61ca5-e248-4027-b705-6d9c5dabe820",
"type": "main",
"index": 0
}
]
]
},
"6d11ae64-e2f8-47ed-854a-c749881ce72c": {
"main": [
[
{
"node": "Google Sheets News Storage",
"type": "main",
"index": 0
}
]
]
}
}
}よくある質問
このワークフローの使い方は?
上記のJSON設定コードをコピーし、n8nインスタンスで新しいワークフローを作成して「JSONからインポート」を選択、設定を貼り付けて認証情報を必要に応じて変更してください。
このワークフローはどんな場面に適していますか?
中級 - 市場調査, AI要約
有料ですか?
このワークフローは完全無料です。ただし、ワークフローで使用するサードパーティサービス(OpenAI APIなど)は別途料金が発生する場合があります。
関連ワークフロー
私のワークフロー 2
ScrapeGraphAIと戦略の計画を用いたAI駆動のコンテンツギャップ分析
Code
Google Sheets
Schedule Trigger
+
Code
Google Sheets
Schedule Trigger
18 ノードvinci-king-01
市場調査
私のワークフロー2
AI競合監視と収益最適化を組み合わせた自動動の管理価格設定
If
Code
Merge
+
If
Code
Merge
25 ノードvinci-king-01
市場調査
私のワークフロー2
ScrapeGraphAIとGoogle Sheetsを使ってSEOのバックリンクを発見・分析する
Code
Filter
Email Send
+
Code
Filter
Email Send
17 ノードvinci-king-01
市場調査
不動産保守コスト分析
ScrapeGraphAIを使って不動産のメンテナンスコストを分析し、予算計画を立てる
Code
Schedule Trigger
Scrapegraph Ai
+
Code
Schedule Trigger
Scrapegraph Ai
12 ノードvinci-king-01
市場調査
不動産市場の感情分析
ScrapeGraphAIとTelegramを使って不動産市場の感情を分析する
Code
Telegram
Schedule Trigger
+
Code
Telegram
Schedule Trigger
15 ノードvinci-king-01
市場調査
財産
ScrapeGraphAIを使用してZillowの不動産一覧をTelegramに自動送信
Code
Telegram
Schedule Trigger
+
Code
Telegram
Schedule Trigger
8 ノードvinci-king-01
市場調査