Datasette
A tool for exploring and publishing data
Datasette is a tool for exploring and publishing data. It helps people take data of any shape or size and publish that as an interactive, explorable website and accompanying API.
Datasette is aimed at data journalists, museum curators, archivists, local governments and anyone else who has data that they wish to share with the world.
Explore a demo, watch a video about the project or try it out by uploading and publishing your own CSV data.
- Comprehensive documentation: http://datasette.readthedocs.io/
- Examples: https://github.com/simonw/datasette/wiki/Datasettes
- Live demo of current master: https://latest.datasette.io/
News
- 29th January 2020: Datasette 0.34 - improvements to search, datasette publish cloudrunanddatasette package.
- 21st January 2020: Deploying a data API using GitHub Actions and Cloud Run - how to use GitHub Actions and Google Cloud Run to automatically scrape data and deploy the result as an API with Datasette.
- 22nd December 2019: Datasette 0.33 - various small improvements.
- 19th December 2019: Building tools to bring data-driven reporting to more newsrooms - some notes on my JSK fellowship so far.
- 2nd December 2019: Niche Museums is a new site entirely powered by Datasette, using custom templates and plugins. niche-museums.com, powered by Datasette describes how the site works, and datasette-atom: Define an Atom feed using a custom SQL query describes how the new datasette-atom plugin was used to add an Atom syndication feed to the site.
- 14th November 2019: Datasette 0.32 now uses asynchronous rendering in Jinja templates, which means template functions can perform asynchronous operations such as executing SQL queries. datasette-template-sql is a new plugin uses this capability to add a new custom sql(sql_query)template function.
- 11th November 2019: Datasette 0.31 - the first version of Datasette to support Python 3.8, which means dropping support for Python 3.5.
- 18th October 2019: Datasette 0.30
- 13th July 2019: Single sign-on against GitHub using ASGI middleware talks about the implementation of datasette-auth-github in more detail.
- 7th July 2019: Datasette 0.29 - ASGI, new plugin hooks, facet by date and much, much more...
- datasette-auth-github - a new plugin for Datasette 0.29 that lets you require users to authenticate against GitHub before accessing your Datasette instance. You can whitelist specific users, or you can restrict access to members of specific GitHub organizations or teams.
- datasette-cors - a plugin that lets you configure CORS access from a list of domains (or a set of domain wildcards) so you can make JavaScript calls to a Datasette instance from a specific set of other hosts.
 
- 23rd June 2019: Porting Datasette to ASGI, and Turtles all the way down
- 21st May 2019: The anonymized raw data from the Stack Overflow Developer Survey 2019 has been published in partnership with Glitch, powered by Datasette.
- 19th May 2019: Datasette 0.28 - a salmagundi of new features!
- No longer immutable! Datasette now supports databases that change.
- Faceting improvements including facet-by-JSON-array and the ability to define custom faceting using plugins.
- datasette publish cloudrun lets you publish databases to Google's new Cloud Run hosting service.
- New register_output_renderer plugin hook for adding custom output extensions to Datasette in addition to the default .jsonand.csv.
- Dozens of other smaller features and tweaks - see the release notes for full details.
- Read more about this release here: Datasette 0.28—and why master should always be releasable
 
- 24th February 2019: 
 sqlite-utils: a Python library and CLI tool for building SQLite databases - a partner tool for easily creating SQLite databases for use with Datasette.
- 31st Janary 2019: Datasette 0.27 - datasette pluginscommand, newline-delimited JSON export option, new documentation on The Datasette Ecosystem.
- 10th January 2019: Datasette 0.26.1 - SQLite upgrade in Docker image, /-/versionsnow shows SQLite compile options.
- 2nd January 2019: Datasette 0.26 - minor bug fixes, datasette publish now --aliasargument.
- 18th December 2018: Fast Autocomplete Search for Your Website - a new tutorial on using Datasette to build a JavaScript autocomplete search engine.
- 3rd October 2018: The interesting ideas in Datasette - a write-up of some of the less obvious interesting ideas embedded in the Datasette project.
- 19th September 2018: Datasette 0.25 - New plugin hooks, improved database view support and an easier way to use more recent versions of SQLite.
- 23rd July 2018: Datasette 0.24 - a number of small new features
- 29th June 2018: datasette-vega, a new plugin for visualizing data as bar, line or scatter charts
- 21st June 2018: Datasette 0.23.1 - minor bug fixes
- 18th June 2018: Datasette 0.23: CSV, SpatiaLite and more - CSV export, foreign key expansion in JSON and CSV, new config options, improved support for SpatiaLite and a bunch of other improvements
- 23rd May 2018: Datasette 0.22.1 bugfix plus we now use versioneer
- 20th May 2018: Datasette 0.22: Datasette Facets
- 5th May 2018: Datasette 0.21: New _shape=, new _size=, search within columns
- 25th April 2018: Exploring the UK Register of Members Interests with SQL and Datasette - a tutorial describing how register-of-members-interests.datasettes.com was built (source code here)
- 20th April 2018: Datasette plugins, and building a clustered map visualization - introducing Datasette's new plugin system and datasette-cluster-map, a plugin for visualizing data on a map
- 20th April 2018: Datasette 0.20: static assets and templates for plugins
- 16th April 2018: Datasette 0.19: plugins preview
- 14th April 2018: Datasette 0.18: units
- 9th April 2018: Datasette 0.15: sort by column
- 28th March 2018: Baltimore Sun Public Salary Records - a data journalism project from the Baltimore Sun powered by Datasette - source code is available here
- 27th March 2018: Cloud-first: Rapid webapp deployment using containers - a tutorial covering deploying Datasette using Microsoft Azure by the Research Software Engineering team at Imperial College London
- 28th January 2018: Analyzing my Twitter followers with Datasette - a tutorial on using Datasette to analyze follower data pulled from the Twitter API
- 17th January 2018: Datasette Publish: a web app for publishing CSV files as an online database
- 12th December 2017: Building a location to time zone API with SpatiaLite, OpenStreetMap and Datasette
- 9th December 2017: Datasette 0.14: customization edition
- 25th November 2017: New in Datasette: filters, foreign keys and search
- 13th November 2017: Datasette: instantly create and publish an API for your SQLite databases
Installation
pip3 install datasette
Datasette requires Python 3.6 or higher. We also have detailed installation instructions covering other options such as Docker.
Basic usage
datasette serve path/to/database.db
This will start a web server on port 8001 - visit http://localhost:8001/ to access the web interface.
serve is the default subcommand, you can omit it if you like.
Use Chrome on OS X? You can run datasette against your browser history like so:
 datasette ~/Library/Application\ Support/Google/Chrome/Default/History
Now visiting http://localhost:8001/History/downloads will show you a web interface to browse your downloads data:

datasette serve options
Usage: datasette serve [OPTIONS] [FILES]...
  Serve up specified SQLite database files with a web UI
Options:
  -i, --immutable PATH      Database files to open in immutable mode
  -h, --host TEXT           Host for server. Defaults to 127.0.0.1 which means
                            only connections from the local machine will be
                            allowed. Use 0.0.0.0 to listen to all IPs and
                            allow access from other machines.
  -p, --port INTEGER        Port for server, defaults to 8001
  --debug                   Enable debug mode - useful for development
  --reload                  Automatically reload if database or code change
                            detected - useful for development
  --cors                    Enable CORS by serving Access-Control-Allow-
                            Origin: *
  --load-extension PATH     Path to a SQLite extension to load
  --inspect-file TEXT       Path to JSON file created using "datasette
                            inspect"
  -m, --metadata FILENAME   Path to JSON file containing license/source
                            metadata
  --template-dir DIRECTORY  Path to directory containing custom templates
  --plugins-dir DIRECTORY   Path to directory containing custom plugins
  --static STATIC MOUNT     mountpoint:path-to-directory for serving static
                            files
  --memory                  Make :memory: database available
  --config CONFIG           Set config option using configname:value
                            datasette.readthedocs.io/en/latest/config.html
  --version-note TEXT       Additional note to show on /-/versions
  --help-config             Show available config options
  --help                    Show this message and exit.
metadata.json
If you want to include licensing and source information in the generated datasette website you can do so using a JSON file that looks something like this:
{
    "title": "Five Thirty Eight",
    "license": "CC Attribution 4.0 License",
    "license_url": "http://creativecommons.org/licenses/by/4.0/",
    "source": "fivethirtyeight/data on GitHub",
    "source_url": "https://github.com/fivethirtyeight/data"
}
Save this in metadata.json and run Datasette like so:
datasette serve fivethirtyeight.db -m metadata.json
The license and source information will be displayed on the index page and in the footer. They will also be included in the JSON produced by the API.
datasette publish
If you have Heroku, Google Cloud Run or Zeit Now v1 configured, Datasette can deploy one or more SQLite databases to the internet with a single command:
datasette publish heroku database.db
Or:
datasette publish cloudrun database.db
This will create a docker image containing both the datasette application and the specified SQLite database files. It will then deploy that image to Heroku or Cloud Run and give you a URL to access the resulting website and API.
See Publishing data in the documentation for more details.
 projecs by owner
                                                                (
                                                                projecs by owner
                                                                (