Interline Transitland
transitland-lib is a library and command-line tool for reading, writing, and processing transit data in GTFS and related formats. The library is structured as a set of data sources, filters, and transformations that can be mixed together in a variety of ways to create processing pipelines. The library supports the DMFR format to describe feed resources.

Table of Contents
Installation
Download prebuilt binary
The transitland binaries for Linux and macOS are attached to each release.
Install using homebrew
The transitland binary can be installed using homebrew. The executable is code-signed and notarized.
brew install interline-io/transitland-lib/transitland-lib
Install binary from source
go get github.com/interline-io/transitland-lib/cmd/transitland
This package uses Go Modules and will also install required dependencies.
Main dependencies:
twpayne/go-geom
jmoiron/sqlx
Masterminds/squirrel
lib/pq
mattn/go-sqlite3 (requires CGO)
The main subcommands are:
validate command
The validate command performs a basic validation on a data source and writes the results to standard out.
% transitland validate --help
Usage: validate <reader>
-ext value
Include GTFS Extension
Example:
% transitland validate "https://www.bart.gov/dev/schedules/google_transit.zip"
copy command
The copy command performs a basic copy from a reader to a writer. By default, any entity with errors will be skipped and not written to output. This can be ignored with -allow-entity-errors to ignore simple errors and -allow-reference-errors to ignore entity relationship errors, such as a reference to a non-existent stop.
% transitland copy --help
Usage: copy <reader> <writer>
-allow-entity-errors
Allow entities with errors to be copied
-allow-reference-errors
Allow entities with reference errors to be copied
-create
Create a basic database schema if none exists
-ext value
Include GTFS Extension
-fvid int
Specify FeedVersionID when writing to a database
Example:
% transitland copy --allow-entity-errors "https://www.bart.gov/dev/schedules/google_transit.zip" output.zip
% unzip -p output.zip agency.txt
agency_id,agency_name,agency_url,agency_timezone,agency_lang,agency_phone,agency_fare_url,agency_email
BART,Bay Area Rapid Transit,https://www.bart.gov/,America/Los_Angeles,,510-464-6000,,
extract command
The extract command extends the basic copy command with a number of additional options and transformations. It can be used to pull out a single route or trip, interpolate stop times, override a single value on an entity, etc. This is a separate command to keep the basic copy command simple while allowing the extract command to grow and add more features over time.
% transitland extract --help
Usage: extract <input> <output>
-allow-entity-errors
Allow entities with errors to be copied
-allow-reference-errors
Allow entities with reference errors to be copied
-create
Create a basic database schema if none exists
-create-missing-shapes
Create missing Shapes from Trip stop-to-stop geometries
-ext value
Include GTFS Extension
-extract-agency value
Extract Agency
-extract-calendar value
Extract Calendar
-extract-route value
Extract Route
-extract-route-type value
Extract Routes matching route_type
-extract-stop value
Extract Stop
-extract-trip value
Extract Trip
-fvid int
Specify FeedVersionID when writing to a database
-interpolate-stop-times
Interpolate missing StopTime arrival/departure values
-normalize-service-ids
Create Calendar entities for CalendarDate service_id's
-set value
Set values on output; format is filename,id,key,value
-use-basic-route-types
Collapse extended route_type's into basic GTFS values
Example:
# Extract a single trip from the BART GTFS, and rename the agency to "test".
% transitland extract -extract-trip "3050453" -set "agency.txt,BART,agency_id,test" "https://www.bart.gov/dev/schedules/google_transit.zip" output2.zip
# Note renamed agency
% unzip -p output2.zip agency.txt
agency_id,agency_name,agency_url,agency_timezone,agency_lang,agency_phone,agency_fare_url,agency_email
test,Bay Area Rapid Transit,https://www.bart.gov/,America/Los_Angeles,,510-464-6000,,
# Only entities related to the specified trip are included in the output.
% unzip -p output2.zip trips.txt
route_id,service_id,trip_id,trip_headsign,trip_short_name,direction_id,block_id,shape_id,wheelchair_accessible,bikes_allowed
1,2020_09_14-DX-MVS-Weekday-15,3050453,San Francisco International Airport,,1,,01_shp,0,0
$ unzip -p output2.zip routes.txt
route_id,agency_id,route_short_name,route_long_name,route_desc,route_type,route_url,route_color,route_text_color,route_sort_order
1,test,YL-S,Antioch to SFIA/Millbrae,,1,http://www.bart.gov/schedules/bylineresults?route=1,FFFF33,,0
% transitland % unzip -p output2.zip stop_times.txt
trip_id,arrival_time,departure_time,stop_id,stop_sequence,stop_headsign,pickup_type,drop_off_type,shape_dist_traveled,timepoint
3050453,04:53:00,04:53:00,CONC,0,,0,0,0.00000,0
3050453,04:58:00,04:58:00,PHIL,2,,0,0,4.06000,0
3050453,05:01:00,05:02:00,WCRK,3,,0,0,5.77000,0
3050453,05:06:00,05:07:00,LAFY,4,,0,0,9.23000,0
3050453,05:11:00,05:12:00,ORIN,5,,0,0,12.99000,0
3050453,05:17:00,05:18:00,ROCK,6,,0,0,17.38000,0
...
dmfr command
under development
The dmfr command enables processing multiple feeds at once using a catalog in the Distributed Mobility Feed Registry format. It provides several additional subcommands for reading DMFR files, synchronizing these feeds to a database, downloading the latest versions of each feed, and automatically importing the feeds into a database. It provides the foundation for Transitland v2.
This command is still under active development and may change in future releases. Please see DMFR Command help.
Usage as a library
Key library components
- Entity: An
Entity is entity as specified by GTFS, such as an Agency, Route, Stop, etc.
- Reader: A
Reader provides streams of GTFS entities over channels. The tlcsv and tldb modules provide CSV and Postgres/SQLite support, respectively.
- Writer: A
Writer accepts GTFS entities. As above, tlcsv and tldb provide basic implementations. Custom writers can also be used to support non-GTFS outputs, such as building a routing graph.
- Copier: A
Copier reads a stream of GTFS entities from a Reader, checks each entity against a Marker, performs validation, applies any specified Filters, and sends to a Writer.
- Marker: A
Marker selects which GTFS entities will be processed by a Copier. For example, selecting only entities related to a single trip or route.
- Filter: A
Filter applies transformations to GTFS entities, such as converting extended route types to basic values, or modifying entity identifiers.
- Extension: An
Extension provides support for additional types of GTFS entities.
See godoc.org for package documentation.
Install as a library
go get github.com/interline-io/transitland-lib
Example of how to use as a library
A simple example of reading and writing GTFS entities from CSV (full example):
package main
import (
"fmt"
"testing"
"github.com/interline-io/transitland-lib/copier"
"github.com/interline-io/transitland-lib/tl"
"github.com/interline-io/transitland-lib/tlcsv"
"github.com/interline-io/transitland-lib/tldb"
)
var URL = "https://github.com/interline-io/transitland-lib/raw/master/test/data/external/bart.zip"
func TestExample1(t *testing.T) {
// Read stops from a GTFS url
reader, _ := tlcsv.NewReader(URL)
reader.Open()
defer reader.Close()
// Write to to the current directory
writer, _ := tlcsv.NewWriter(".")
writer.Open()
// Copy stops
for stop := range reader.Stops() {
fmt.Println("Read Stop:", stop.StopID)
eid, _ := writer.AddEntity(&stop)
fmt.Println("Wrote stop:", eid)
}
}
Database support is handled similary:
func getReader() tl.Reader {
reader, _ := tlcsv.NewReader(URL)
return reader
}
func TestExample2(t *testing.T) {
reader := getReader()
// Create a SQLite writer, in memory
dburl := "sqlite3://:memory:"
dbwriter, err := tldb.NewWriter(dburl)
if err != nil {
t.Fatalf("no reader available")
}
if err := dbwriter.Open(); err != nil {
t.Fatalf("could not open database for writing")
}
if err := dbwriter.Create(); err != nil {
t.Fatalf("could not find or create database schema")
}
for stop := range reader.Stops() {
// A database writer AddEntity returns the primary key as a string.
fmt.Println("Read Stop:", stop.StopID)
eid, err := dbwriter.AddEntity(&stop)
if err != nil {
t.Fatalf("could not write entity to database")
}
fmt.Println("wrote stop to database:", eid)
}
// Read back from this source.
dbreader, err := dbwriter.NewReader()
if err != nil {
t.Fatalf("could not get a new reader")
}
count := 0
for stop := range dbreader.Stops() {
fmt.Println("read stop from database:", stop.StopID)
count++
}
if count != 50 {
t.Errorf("got %d stops, expected 50", count)
}
}
More advanced filtering operations can be performed using a Copier, which provides additional hooks for filtering, transformation, and validation:
func TestExample3(t *testing.T) {
reader := getReader()
// Create a zip writer
writer, err := tlcsv.NewWriter("filtered.zip")
if err != nil {
t.Fatalf("no writer available")
}
// Create a copier to stream, filter, and validate entities
cp := copier.NewCopier(reader, writer)
result := cp.Copy()
if result.WriteError != nil {
t.Fatalf("fatal copy error")
}
for _, err := range result.Errors {
fmt.Println("error:", err)
}
for fn, count := range result.EntityCount {
fmt.Printf("copied %d entities from %s\n", count, fn)
}
}
See API docs at https://godoc.org/github.com/interline-io/transitland-lib
Usage as a Web Service
transitland-lib can be used in a variety of ways to power a web service. Interline currently uses two approaches:
-
Populate a database with one or more feeds using transitland-lib and use the transitland server command to serve the Transitland v2 REST and/or v2 GraphQL API endpoints. These API endpoints are primarily read-only and focused on querying and analyzing transit data.
-
Populate a Postgres database with one or more feeds using transitland-lib, or just create an empty database using transitland-lib's schema. Use Hasura to provide a complete GraphQL API for reading and writing into the database.
For more information about how these web services are used within the overall architecture of the Transitland platform, see https://www.transit.land/documentation#transitland-architecture
server command
To start the server with the REST API endpoints, GraphQL API endpoint, GraphQL explorer UI, image generation endpoints, and without authentication (all requests run as admin):
transitland server -playground -auth=admin -dburl=postgres://localhost/transitland_db?sslmode=disable
The above command assumes you have a Postgres database named "transitland_db" running on your machine under the default Postgres port. Alternatively set the database URL using the TL_DATABASE_URL environment variable.
Open http://localhost:8080/ in your web browser or query it using an HTTP client.
To start stripped down, with only REST API endpoints (no GraphQL API or explorer UI, no static image API endpoints) and with JWT authentication:
transitland server -disable-image -disable-graphql -auth=jwt -jwt-audience="http://example.com" -jwt-issuer="<issuer_id>" -jwt-public-key-file="<key.pem>"
Note that if you are using transitland server with JWT authentication, place the process behind a reverse-proxy or gateway that can issue JWT tokens. At Interline, we use the combination of Kong and Auth0.
More options:
% transitland server --help
Usage: server
-auth string
-dburl string
Database URL (default: $TL_DATABASE_URL)
-disable-graphql
Disable GraphQL endpoint
-disable-image
Disable image generation
-disable-rest
Disable REST endpoint
-gtfsdir string
Directory to store GTFS files
-jwt-audience string
JWT Audience
-jwt-issuer string
JWT Issuer
-jwt-public-key-file string
Path to JWT public key file
-playground
Enable GraphQL playground
-port string
(default "8080")
-s3 string
S3 bucket for GTFS files
-timeout int
(default 60)
-validate-large-files
Allow validation of large files
Hasura
Hasura is a web service that can provide an "instant" GraphQL API based on a Postgres database and its schema. We combine Hasura with transitland-lib for projects that involve creating new or complex queries (since Hasura can be more flexible than the queries provided by transitland server) and projects that involve an API with full read and write access (for example, editing GTFS data, which is also not provided by transitland server). Note that Hasura's automatically generated database queries are not guaranteed to be efficient (on the other hand, transitland server is tuned to provide better performance).
To use Hasura with transitland-lib you can either import feeds into a new Postgres database (using the transitland dmfr command) or create a blank Postgres database (using the schema in internal/schema/postgres.pgsql). Configure Hasura to recognize all the tables and the foreign key relationships between them.
Included Readers and Writers
| Target |
Module |
Supports Read |
Supports Write |
| CSV |
tlcsv |
✅ |
✅ |
| SQLite |
tldb |
✅ |
✅ |
| Postgres (with PostGIS) |
tldb |
✅ |
✅ |
We welcome the addition of more readers and writers.
Development
transitland-lib follows Go coding conventions.
GitHub Actions runs all tests, stores code coverage reports as artifacts, and prepares releases.
Releases
Releases follow Semantic Versioning conventions.
To cut a new release:
- Update
transitland-lib/tl/tl.go with the new version.
- Create a GitHub release. This will create a tag and GitHub Actions will create & attach code-signed binaries.
- Download the files from the release, and update the homebrew formula with the updated sha256 hashes and version tag.
Licenses
transitland-lib is released under a "dual license" model:
- open-source for use by all under the GPLv3 license
- also available under a flexible commercial license from Interline