HN Who is Hiring Scrape

用于HN Who is Hiring Scrape,自动化提取和整理Hacker News上“谁在招聘”的帖子,提供最新的招聘信息,确保数据结构统一,便于分析和存储。通过集成Algolia和OpenAI,快速获取和清洗数据,并将结果写入Airtable,提升招聘信息的可访问性和可用性。

7/8/2025
20 nodes
Complex
manualcomplexsticky notesplitoutlangchainfilterairtableadvancedapiintegration
Categories:
Complex WorkflowManual TriggeredData Processing & Analysis
Integrations:
Sticky NoteSplitOutLangChainFilterAirtable

Target Audience

  • Job Seekers: Individuals looking for job opportunities in the tech industry, particularly those interested in roles advertised on Hacker News.
    - Recruiters: Professionals seeking to gather information about companies that are actively hiring and the roles they are offering.
    - Developers and Data Enthusiasts: Those interested in automating data collection processes and leveraging APIs for job market insights.
    - Entrepreneurs and Startups: Business owners looking to understand hiring trends and opportunities in their industry.
  • Problem Solved

    This workflow addresses the challenge of efficiently gathering job postings from Hacker News, particularly the monthly 'Who is Hiring?' thread. It automates the retrieval of job listings and their details, allowing users to access this information without manual searching. By utilizing the Hacker News API and Algolia search, users can quickly obtain the latest job opportunities along with relevant details such as job title, company, location, and application links.

    Workflow Steps

  • Trigger the Workflow: The process begins when the user manually triggers the workflow.
    - Search for Job Posts: It sends a request to the Algolia API to retrieve posts from the 'Ask HN: Who is hiring?' thread, filtering and sorting them by date.
    - Extract Relevant Data: The workflow extracts essential information such as title, createdAt, and storyId from the retrieved posts.
    - Get Latest Post: It filters the results to focus on posts created within the last 30 days.
    - Fetch Individual Job Details: For each job post, it retrieves additional details using the Hacker News API.
    - Clean and Structure Data: The text from the job posts is cleaned and structured into a consistent format using OpenAI's language model.
    - Output to Airtable: Finally, the structured job data is written to an Airtable base for easy access and management.
  • Customization Guide

  • API Credentials: Users need to set up their own Algolia and Hacker News API credentials to ensure the workflow functions correctly.
    - Modify Search Parameters: Users can adjust the search query or filters in the HTTP request to target specific job types or companies.
    - Change Output Destination: Instead of Airtable, users can modify the final node to send data to other platforms, such as Google Sheets or a custom database.
    - Enhance Data Processing: Users can customize the text cleaning and structuring logic in the code node to better fit their specific data requirements or formats.