Initial cut at datasette publish heroku

Rather gross, but proves that it works.
This commit is contained in:
Jacob Kaplan-Moss 2017-11-15 11:53:00 -08:00
commit f48cb705d8
2 changed files with 77 additions and 19 deletions

View file

@ -7,6 +7,7 @@ import sqlite3
import tempfile
import time
import urllib
import shlex
def compound_pks_from_path(path):
@ -197,3 +198,48 @@ def temporary_docker_directory(files, name, metadata, extra_options, extra_metad
finally:
tmp.cleanup()
os.chdir(saved_cwd)
@contextmanager
def temporary_heroku_directory(files, name, metadata, extra_options, extra_metadata=None):
# FIXME: lots of duplicated code from above
extra_metadata = extra_metadata or {}
tmp = tempfile.TemporaryDirectory()
saved_cwd = os.getcwd()
file_paths = [
os.path.join(saved_cwd, name)
for name in files
]
file_names = [os.path.split(f)[-1] for f in files]
if metadata:
metadata_content = json.load(metadata)
else:
metadata_content = {}
for key, value in extra_metadata.items():
if value:
metadata_content[key] = value
try:
os.chdir(tmp.name)
if metadata_content:
open('metadata.json', 'w').write(json.dumps(metadata_content, indent=2))
open('runtime.txt', 'w').write('python-3.6.2')
open('requirements.txt', 'w').write('datasette')
quoted_files = " ".join(map(shlex.quote, files))
procfile_cmd = f'web: datasette serve --host 0.0.0.0 {quoted_files} --cors --port $PORT --inspect-file inspect-data.json'
open('Procfile', 'w').write(procfile_cmd)
for path, filename in zip(file_paths, file_names):
os.link(path, os.path.join(tmp.name, filename))
yield
finally:
tmp.cleanup()
os.chdir(saved_cwd)