poll-scraper

command module
v1.0.1 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Feb 20, 2026 License: BSD-3-Clause Imports: 19 Imported by: 0

README

Poll Scraper

A desktop application designed to collect and structure data from the savitarna.mobi.lt website with the purpose of enabling its integration into various graphic engines.


Features

Scraper Control

Start and stop the scraper from the UI. The scraper fetches data from configured URLs at a set interval and writes results to output files. Use Preview to fetch data once without starting the continuous scraper.

Filter Lines

Choose which scraped lines to include in the output. The filter modal shows all lines grouped by their source URL, so you can easily identify where each line comes from. Filters use stable indices tied to URL data only, so adding or removing custom lines won't shift your selections.

Custom Lines

Add extra data rows with a name and value that get appended after the scraped data. Each custom line has a visibility toggle — hidden lines are excluded from output without deleting them.

Sum

Automatically calculates the sum of all numeric values in the output. Optionally append a symbol (e.g. $, ) after the sum value. Adds one or two extra lines at the end: sum and optionally sum_symbol.

Line Count Protection

If any URL starts returning a different number of lines than the first scrape, an error is logged and the scraper status becomes faulted. Optionally enable Stop on URL line count change in settings to automatically stop the scraper when this happens.


Output

TXT

Writes data in a structured format for Textus Live. Choose between ANSI (Windows-1252) or UTF-8 encoding.

CSV

Writes name/value pairs as CSV rows.

HTTP API

JSON API. Any client can fetch the current data as a JSON array from the root endpoint. CORS domains can be restricted in settings.


UI

Status Tab

Shows scraper state, URL health (which URLs are returning data), line counts, output targets, and server status. Includes a live log viewer with level filtering (Error, Warn, Info, Debug) and color coding.

Settings Tab

Update interval, debug mode, line count protection, server configuration, and output paths with encoding selection.

Scraping Tab

URL list, data parsing options, custom lines management, and line filter configuration.

Status Indicator
Color Meaning
Green Scraper is running, all URLs healthy
Gray Scraper is stopped
Red Error — URL failures, write errors, or line count changes

Setup

  1. Install Go 1.21+ and Wails
  2. Run: wails dev

A default config.json is created automatically on first launch if one doesn't exist.


Build

wails build

Notes

  • All settings are configurable through the UI — the config.json file is managed automatically
  • Errors are logged to error.log

Documentation

The Go Gopher

There is no documentation for this package.

Directories

Path Synopsis
api

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL