mirror of
https://github.com/simonw/datasette.git
synced 2025-12-10 16:51:24 +01:00
Facet by date, closes #481
This commit is contained in:
parent
2d14f21f85
commit
967230c90e
6 changed files with 231 additions and 40 deletions
|
|
@ -30,7 +30,6 @@ def load_facet_configs(request, table_metadata):
|
||||||
type = "column"
|
type = "column"
|
||||||
metadata_config = {"simple": metadata_config}
|
metadata_config = {"simple": metadata_config}
|
||||||
else:
|
else:
|
||||||
# This should have a single key and a single value
|
|
||||||
assert (
|
assert (
|
||||||
len(metadata_config.values()) == 1
|
len(metadata_config.values()) == 1
|
||||||
), "Metadata config dicts should be {type: config}"
|
), "Metadata config dicts should be {type: config}"
|
||||||
|
|
@ -62,7 +61,7 @@ def load_facet_configs(request, table_metadata):
|
||||||
|
|
||||||
@hookimpl
|
@hookimpl
|
||||||
def register_facet_classes():
|
def register_facet_classes():
|
||||||
classes = [ColumnFacet]
|
classes = [ColumnFacet, DateFacet]
|
||||||
if detect_json1():
|
if detect_json1():
|
||||||
classes.append(ArrayFacet)
|
classes.append(ArrayFacet)
|
||||||
return classes
|
return classes
|
||||||
|
|
@ -365,3 +364,116 @@ class ArrayFacet(Facet):
|
||||||
facets_timed_out.append(column)
|
facets_timed_out.append(column)
|
||||||
|
|
||||||
return facet_results, facets_timed_out
|
return facet_results, facets_timed_out
|
||||||
|
|
||||||
|
|
||||||
|
class DateFacet(Facet):
|
||||||
|
type = "date"
|
||||||
|
|
||||||
|
async def suggest(self):
|
||||||
|
columns = await self.get_columns(self.sql, self.params)
|
||||||
|
already_enabled = [c["config"]["simple"] for c in self.get_configs()]
|
||||||
|
suggested_facets = []
|
||||||
|
for column in columns:
|
||||||
|
if column in already_enabled:
|
||||||
|
continue
|
||||||
|
# Does this column contain any dates in the first 100 rows?
|
||||||
|
suggested_facet_sql = """
|
||||||
|
select date({column}) from (
|
||||||
|
{sql}
|
||||||
|
) where {column} glob "????-??-*" limit 100;
|
||||||
|
""".format(
|
||||||
|
column=escape_sqlite(column), sql=self.sql
|
||||||
|
)
|
||||||
|
try:
|
||||||
|
results = await self.ds.execute(
|
||||||
|
self.database,
|
||||||
|
suggested_facet_sql,
|
||||||
|
self.params,
|
||||||
|
truncate=False,
|
||||||
|
custom_time_limit=self.ds.config("facet_suggest_time_limit_ms"),
|
||||||
|
log_sql_errors=False,
|
||||||
|
)
|
||||||
|
values = tuple(r[0] for r in results.rows)
|
||||||
|
if any(values):
|
||||||
|
suggested_facets.append(
|
||||||
|
{
|
||||||
|
"name": column,
|
||||||
|
"type": "date",
|
||||||
|
"toggle_url": self.ds.absolute_url(
|
||||||
|
self.request,
|
||||||
|
path_with_added_args(
|
||||||
|
self.request, {"_facet_date": column}
|
||||||
|
),
|
||||||
|
),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
except (InterruptedError, sqlite3.OperationalError):
|
||||||
|
continue
|
||||||
|
return suggested_facets
|
||||||
|
|
||||||
|
async def facet_results(self):
|
||||||
|
facet_results = {}
|
||||||
|
facets_timed_out = []
|
||||||
|
args = dict(self.get_querystring_pairs())
|
||||||
|
facet_size = self.ds.config("default_facet_size")
|
||||||
|
for source_and_config in self.get_configs():
|
||||||
|
config = source_and_config["config"]
|
||||||
|
source = source_and_config["source"]
|
||||||
|
column = config.get("column") or config["simple"]
|
||||||
|
# TODO: does this query break if inner sql produces value or count columns?
|
||||||
|
facet_sql = """
|
||||||
|
select date({col}) as value, count(*) as count from (
|
||||||
|
{sql}
|
||||||
|
)
|
||||||
|
where date({col}) is not null
|
||||||
|
group by date({col}) order by count desc limit {limit}
|
||||||
|
""".format(
|
||||||
|
col=escape_sqlite(column), sql=self.sql, limit=facet_size + 1
|
||||||
|
)
|
||||||
|
try:
|
||||||
|
facet_rows_results = await self.ds.execute(
|
||||||
|
self.database,
|
||||||
|
facet_sql,
|
||||||
|
self.params,
|
||||||
|
truncate=False,
|
||||||
|
custom_time_limit=self.ds.config("facet_time_limit_ms"),
|
||||||
|
)
|
||||||
|
facet_results_values = []
|
||||||
|
facet_results[column] = {
|
||||||
|
"name": column,
|
||||||
|
"type": self.type,
|
||||||
|
"results": facet_results_values,
|
||||||
|
"hideable": source != "metadata",
|
||||||
|
"toggle_url": path_with_removed_args(
|
||||||
|
self.request, {"_facet_date": column}
|
||||||
|
),
|
||||||
|
"truncated": len(facet_rows_results) > facet_size,
|
||||||
|
}
|
||||||
|
facet_rows = facet_rows_results.rows[:facet_size]
|
||||||
|
for row in facet_rows:
|
||||||
|
selected = str(args.get("{}__date".format(column))) == str(
|
||||||
|
row["value"]
|
||||||
|
)
|
||||||
|
if selected:
|
||||||
|
toggle_path = path_with_removed_args(
|
||||||
|
self.request, {"{}__date".format(column): str(row["value"])}
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
toggle_path = path_with_added_args(
|
||||||
|
self.request, {"{}__date".format(column): row["value"]}
|
||||||
|
)
|
||||||
|
facet_results_values.append(
|
||||||
|
{
|
||||||
|
"value": row["value"],
|
||||||
|
"label": row["value"],
|
||||||
|
"count": row["count"],
|
||||||
|
"toggle_url": self.ds.absolute_url(
|
||||||
|
self.request, toggle_path
|
||||||
|
),
|
||||||
|
"selected": selected,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
except InterruptedError:
|
||||||
|
facets_timed_out.append(column)
|
||||||
|
|
||||||
|
return facet_results, facets_timed_out
|
||||||
|
|
|
||||||
|
|
@ -138,4 +138,11 @@ If your SQLite installation provides the ``json1`` extension (you can check usin
|
||||||
|
|
||||||
This is useful for modelling things like tags without needing to break them out into a new table.
|
This is useful for modelling things like tags without needing to break them out into a new table.
|
||||||
|
|
||||||
You can try this functionality out at `latest.datasette.io/fixtures/facetable?_facet_array=tags <https://latest.datasette.io/fixtures/facetable?_facet_array=tags>`__
|
You can try this functionality out at `latest.datasette.io/fixtures/facetable?_facet_array=tags <https://latest.datasette.io/fixtures/facetable?_facet_array=tags>`__
|
||||||
|
|
||||||
|
Facet by date
|
||||||
|
-------------
|
||||||
|
|
||||||
|
If Datasette finds any columns that contain dates in the first 100 values, it will offer a faceting interface against the dates of those values. This works especially well against timestamp values such as ``2019-03-01 12:44:00``.
|
||||||
|
|
||||||
|
Demo here: `latest.datasette.io/fixtures/facetable?_facet_date=created <https://latest.datasette.io/fixtures/facetable?_facet_date=created>`__
|
||||||
|
|
|
||||||
|
|
@ -518,6 +518,7 @@ INSERT INTO facet_cities (id, name) VALUES
|
||||||
|
|
||||||
CREATE TABLE facetable (
|
CREATE TABLE facetable (
|
||||||
pk integer primary key,
|
pk integer primary key,
|
||||||
|
created text,
|
||||||
planet_int integer,
|
planet_int integer,
|
||||||
on_earth integer,
|
on_earth integer,
|
||||||
state text,
|
state text,
|
||||||
|
|
@ -527,23 +528,23 @@ CREATE TABLE facetable (
|
||||||
FOREIGN KEY ("city_id") REFERENCES [facet_cities](id)
|
FOREIGN KEY ("city_id") REFERENCES [facet_cities](id)
|
||||||
);
|
);
|
||||||
INSERT INTO facetable
|
INSERT INTO facetable
|
||||||
(planet_int, on_earth, state, city_id, neighborhood, tags)
|
(created, planet_int, on_earth, state, city_id, neighborhood, tags)
|
||||||
VALUES
|
VALUES
|
||||||
(1, 1, 'CA', 1, 'Mission', '["tag1", "tag2"]'),
|
("2019-01-14 08:00:00", 1, 1, 'CA', 1, 'Mission', '["tag1", "tag2"]'),
|
||||||
(1, 1, 'CA', 1, 'Dogpatch', '["tag1", "tag3"]'),
|
("2019-01-14 08:00:00", 1, 1, 'CA', 1, 'Dogpatch', '["tag1", "tag3"]'),
|
||||||
(1, 1, 'CA', 1, 'SOMA', '[]'),
|
("2019-01-14 08:00:00", 1, 1, 'CA', 1, 'SOMA', '[]'),
|
||||||
(1, 1, 'CA', 1, 'Tenderloin', '[]'),
|
("2019-01-14 08:00:00", 1, 1, 'CA', 1, 'Tenderloin', '[]'),
|
||||||
(1, 1, 'CA', 1, 'Bernal Heights', '[]'),
|
("2019-01-15 08:00:00", 1, 1, 'CA', 1, 'Bernal Heights', '[]'),
|
||||||
(1, 1, 'CA', 1, 'Hayes Valley', '[]'),
|
("2019-01-15 08:00:00", 1, 1, 'CA', 1, 'Hayes Valley', '[]'),
|
||||||
(1, 1, 'CA', 2, 'Hollywood', '[]'),
|
("2019-01-15 08:00:00", 1, 1, 'CA', 2, 'Hollywood', '[]'),
|
||||||
(1, 1, 'CA', 2, 'Downtown', '[]'),
|
("2019-01-15 08:00:00", 1, 1, 'CA', 2, 'Downtown', '[]'),
|
||||||
(1, 1, 'CA', 2, 'Los Feliz', '[]'),
|
("2019-01-16 08:00:00", 1, 1, 'CA', 2, 'Los Feliz', '[]'),
|
||||||
(1, 1, 'CA', 2, 'Koreatown', '[]'),
|
("2019-01-16 08:00:00", 1, 1, 'CA', 2, 'Koreatown', '[]'),
|
||||||
(1, 1, 'MI', 3, 'Downtown', '[]'),
|
("2019-01-16 08:00:00", 1, 1, 'MI', 3, 'Downtown', '[]'),
|
||||||
(1, 1, 'MI', 3, 'Greektown', '[]'),
|
("2019-01-17 08:00:00", 1, 1, 'MI', 3, 'Greektown', '[]'),
|
||||||
(1, 1, 'MI', 3, 'Corktown', '[]'),
|
("2019-01-17 08:00:00", 1, 1, 'MI', 3, 'Corktown', '[]'),
|
||||||
(1, 1, 'MI', 3, 'Mexicantown', '[]'),
|
("2019-01-17 08:00:00", 1, 1, 'MI', 3, 'Mexicantown', '[]'),
|
||||||
(2, 0, 'MC', 4, 'Arcadia Planitia', '[]')
|
("2019-01-17 08:00:00", 2, 0, 'MC', 4, 'Arcadia Planitia', '[]')
|
||||||
;
|
;
|
||||||
|
|
||||||
CREATE TABLE binary_data (
|
CREATE TABLE binary_data (
|
||||||
|
|
|
||||||
|
|
@ -167,6 +167,7 @@ def test_database_page(app_client):
|
||||||
{
|
{
|
||||||
"columns": [
|
"columns": [
|
||||||
"pk",
|
"pk",
|
||||||
|
"created",
|
||||||
"planet_int",
|
"planet_int",
|
||||||
"on_earth",
|
"on_earth",
|
||||||
"state",
|
"state",
|
||||||
|
|
@ -955,14 +956,16 @@ def test_table_filter_queries_multiple_of_same_type(app_client):
|
||||||
def test_table_filter_json_arraycontains(app_client):
|
def test_table_filter_json_arraycontains(app_client):
|
||||||
response = app_client.get("/fixtures/facetable.json?tags__arraycontains=tag1")
|
response = app_client.get("/fixtures/facetable.json?tags__arraycontains=tag1")
|
||||||
assert [
|
assert [
|
||||||
[1, 1, 1, "CA", 1, "Mission", '["tag1", "tag2"]'],
|
[1, "2019-01-14 08:00:00", 1, 1, "CA", 1, "Mission", '["tag1", "tag2"]'],
|
||||||
[2, 1, 1, "CA", 1, "Dogpatch", '["tag1", "tag3"]'],
|
[2, "2019-01-14 08:00:00", 1, 1, "CA", 1, "Dogpatch", '["tag1", "tag3"]'],
|
||||||
] == response.json["rows"]
|
] == response.json["rows"]
|
||||||
|
|
||||||
|
|
||||||
def test_table_filter_extra_where(app_client):
|
def test_table_filter_extra_where(app_client):
|
||||||
response = app_client.get("/fixtures/facetable.json?_where=neighborhood='Dogpatch'")
|
response = app_client.get("/fixtures/facetable.json?_where=neighborhood='Dogpatch'")
|
||||||
assert [[2, 1, 1, "CA", 1, "Dogpatch", '["tag1", "tag3"]']] == response.json["rows"]
|
assert [
|
||||||
|
[2, "2019-01-14 08:00:00", 1, 1, "CA", 1, "Dogpatch", '["tag1", "tag3"]']
|
||||||
|
] == response.json["rows"]
|
||||||
|
|
||||||
|
|
||||||
def test_table_filter_extra_where_invalid(app_client):
|
def test_table_filter_extra_where_invalid(app_client):
|
||||||
|
|
@ -1331,12 +1334,14 @@ def test_suggested_facets(app_client):
|
||||||
]
|
]
|
||||||
]
|
]
|
||||||
expected = [
|
expected = [
|
||||||
|
{"name": "created", "querystring": "_facet=created"},
|
||||||
{"name": "planet_int", "querystring": "_facet=planet_int"},
|
{"name": "planet_int", "querystring": "_facet=planet_int"},
|
||||||
{"name": "on_earth", "querystring": "_facet=on_earth"},
|
{"name": "on_earth", "querystring": "_facet=on_earth"},
|
||||||
{"name": "state", "querystring": "_facet=state"},
|
{"name": "state", "querystring": "_facet=state"},
|
||||||
{"name": "city_id", "querystring": "_facet=city_id"},
|
{"name": "city_id", "querystring": "_facet=city_id"},
|
||||||
{"name": "neighborhood", "querystring": "_facet=neighborhood"},
|
{"name": "neighborhood", "querystring": "_facet=neighborhood"},
|
||||||
{"name": "tags", "querystring": "_facet=tags"},
|
{"name": "tags", "querystring": "_facet=tags"},
|
||||||
|
{"name": "created", "querystring": "_facet_date=created"},
|
||||||
]
|
]
|
||||||
if detect_json1():
|
if detect_json1():
|
||||||
expected.append({"name": "tags", "querystring": "_facet_array=tags"})
|
expected.append({"name": "tags", "querystring": "_facet_array=tags"})
|
||||||
|
|
@ -1364,6 +1369,7 @@ def test_expand_labels(app_client):
|
||||||
assert {
|
assert {
|
||||||
"2": {
|
"2": {
|
||||||
"pk": 2,
|
"pk": 2,
|
||||||
|
"created": "2019-01-14 08:00:00",
|
||||||
"planet_int": 1,
|
"planet_int": 1,
|
||||||
"on_earth": 1,
|
"on_earth": 1,
|
||||||
"state": "CA",
|
"state": "CA",
|
||||||
|
|
@ -1373,6 +1379,7 @@ def test_expand_labels(app_client):
|
||||||
},
|
},
|
||||||
"13": {
|
"13": {
|
||||||
"pk": 13,
|
"pk": 13,
|
||||||
|
"created": "2019-01-17 08:00:00",
|
||||||
"planet_int": 1,
|
"planet_int": 1,
|
||||||
"on_earth": 1,
|
"on_earth": 1,
|
||||||
"state": "MI",
|
"state": "MI",
|
||||||
|
|
|
||||||
|
|
@ -21,22 +21,22 @@ world
|
||||||
)
|
)
|
||||||
|
|
||||||
EXPECTED_TABLE_WITH_LABELS_CSV = """
|
EXPECTED_TABLE_WITH_LABELS_CSV = """
|
||||||
pk,planet_int,on_earth,state,city_id,city_id_label,neighborhood,tags
|
pk,created,planet_int,on_earth,state,city_id,city_id_label,neighborhood,tags
|
||||||
1,1,1,CA,1,San Francisco,Mission,"[""tag1"", ""tag2""]"
|
1,2019-01-14 08:00:00,1,1,CA,1,San Francisco,Mission,"[""tag1"", ""tag2""]"
|
||||||
2,1,1,CA,1,San Francisco,Dogpatch,"[""tag1"", ""tag3""]"
|
2,2019-01-14 08:00:00,1,1,CA,1,San Francisco,Dogpatch,"[""tag1"", ""tag3""]"
|
||||||
3,1,1,CA,1,San Francisco,SOMA,[]
|
3,2019-01-14 08:00:00,1,1,CA,1,San Francisco,SOMA,[]
|
||||||
4,1,1,CA,1,San Francisco,Tenderloin,[]
|
4,2019-01-14 08:00:00,1,1,CA,1,San Francisco,Tenderloin,[]
|
||||||
5,1,1,CA,1,San Francisco,Bernal Heights,[]
|
5,2019-01-15 08:00:00,1,1,CA,1,San Francisco,Bernal Heights,[]
|
||||||
6,1,1,CA,1,San Francisco,Hayes Valley,[]
|
6,2019-01-15 08:00:00,1,1,CA,1,San Francisco,Hayes Valley,[]
|
||||||
7,1,1,CA,2,Los Angeles,Hollywood,[]
|
7,2019-01-15 08:00:00,1,1,CA,2,Los Angeles,Hollywood,[]
|
||||||
8,1,1,CA,2,Los Angeles,Downtown,[]
|
8,2019-01-15 08:00:00,1,1,CA,2,Los Angeles,Downtown,[]
|
||||||
9,1,1,CA,2,Los Angeles,Los Feliz,[]
|
9,2019-01-16 08:00:00,1,1,CA,2,Los Angeles,Los Feliz,[]
|
||||||
10,1,1,CA,2,Los Angeles,Koreatown,[]
|
10,2019-01-16 08:00:00,1,1,CA,2,Los Angeles,Koreatown,[]
|
||||||
11,1,1,MI,3,Detroit,Downtown,[]
|
11,2019-01-16 08:00:00,1,1,MI,3,Detroit,Downtown,[]
|
||||||
12,1,1,MI,3,Detroit,Greektown,[]
|
12,2019-01-17 08:00:00,1,1,MI,3,Detroit,Greektown,[]
|
||||||
13,1,1,MI,3,Detroit,Corktown,[]
|
13,2019-01-17 08:00:00,1,1,MI,3,Detroit,Corktown,[]
|
||||||
14,1,1,MI,3,Detroit,Mexicantown,[]
|
14,2019-01-17 08:00:00,1,1,MI,3,Detroit,Mexicantown,[]
|
||||||
15,2,0,MC,4,Memnonia,Arcadia Planitia,[]
|
15,2019-01-17 08:00:00,2,0,MC,4,Memnonia,Arcadia Planitia,[]
|
||||||
""".lstrip().replace(
|
""".lstrip().replace(
|
||||||
"\n", "\r\n"
|
"\n", "\r\n"
|
||||||
)
|
)
|
||||||
|
|
|
||||||
|
|
@ -1,4 +1,4 @@
|
||||||
from datasette.facets import ColumnFacet, ArrayFacet
|
from datasette.facets import ColumnFacet, ArrayFacet, DateFacet
|
||||||
from datasette.utils import detect_json1
|
from datasette.utils import detect_json1
|
||||||
from .fixtures import app_client # noqa
|
from .fixtures import app_client # noqa
|
||||||
from .utils import MockRequest
|
from .utils import MockRequest
|
||||||
|
|
@ -17,6 +17,7 @@ async def test_column_facet_suggest(app_client):
|
||||||
)
|
)
|
||||||
suggestions = await facet.suggest()
|
suggestions = await facet.suggest()
|
||||||
assert [
|
assert [
|
||||||
|
{"name": "created", "toggle_url": "http://localhost/?_facet=created"},
|
||||||
{"name": "planet_int", "toggle_url": "http://localhost/?_facet=planet_int"},
|
{"name": "planet_int", "toggle_url": "http://localhost/?_facet=planet_int"},
|
||||||
{"name": "on_earth", "toggle_url": "http://localhost/?_facet=on_earth"},
|
{"name": "on_earth", "toggle_url": "http://localhost/?_facet=on_earth"},
|
||||||
{"name": "state", "toggle_url": "http://localhost/?_facet=state"},
|
{"name": "state", "toggle_url": "http://localhost/?_facet=state"},
|
||||||
|
|
@ -37,6 +38,10 @@ async def test_column_facet_suggest_skip_if_already_selected(app_client):
|
||||||
)
|
)
|
||||||
suggestions = await facet.suggest()
|
suggestions = await facet.suggest()
|
||||||
assert [
|
assert [
|
||||||
|
{
|
||||||
|
"name": "created",
|
||||||
|
"toggle_url": "http://localhost/?_facet=planet_int&_facet=on_earth&_facet=created",
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"name": "state",
|
"name": "state",
|
||||||
"toggle_url": "http://localhost/?_facet=planet_int&_facet=on_earth&_facet=state",
|
"toggle_url": "http://localhost/?_facet=planet_int&_facet=on_earth&_facet=state",
|
||||||
|
|
@ -67,7 +72,14 @@ async def test_column_facet_suggest_skip_if_enabled_by_metadata(app_client):
|
||||||
metadata={"facets": ["city_id"]},
|
metadata={"facets": ["city_id"]},
|
||||||
)
|
)
|
||||||
suggestions = [s["name"] for s in await facet.suggest()]
|
suggestions = [s["name"] for s in await facet.suggest()]
|
||||||
assert ["planet_int", "on_earth", "state", "neighborhood", "tags"] == suggestions
|
assert [
|
||||||
|
"created",
|
||||||
|
"planet_int",
|
||||||
|
"on_earth",
|
||||||
|
"state",
|
||||||
|
"neighborhood",
|
||||||
|
"tags",
|
||||||
|
] == suggestions
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.asyncio
|
@pytest.mark.asyncio
|
||||||
|
|
@ -239,3 +251,55 @@ async def test_array_facet_results(app_client):
|
||||||
"truncated": False,
|
"truncated": False,
|
||||||
}
|
}
|
||||||
} == buckets
|
} == buckets
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.asyncio
|
||||||
|
async def test_date_facet_results(app_client):
|
||||||
|
facet = DateFacet(
|
||||||
|
app_client.ds,
|
||||||
|
MockRequest("http://localhost/?_facet_date=created"),
|
||||||
|
database="fixtures",
|
||||||
|
sql="select * from facetable",
|
||||||
|
table="facetable",
|
||||||
|
)
|
||||||
|
buckets, timed_out = await facet.facet_results()
|
||||||
|
assert [] == timed_out
|
||||||
|
assert {
|
||||||
|
"created": {
|
||||||
|
"name": "created",
|
||||||
|
"type": "date",
|
||||||
|
"results": [
|
||||||
|
{
|
||||||
|
"value": "2019-01-14",
|
||||||
|
"label": "2019-01-14",
|
||||||
|
"count": 4,
|
||||||
|
"toggle_url": "http://localhost/?_facet_date=created&created__date=2019-01-14",
|
||||||
|
"selected": False,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"value": "2019-01-15",
|
||||||
|
"label": "2019-01-15",
|
||||||
|
"count": 4,
|
||||||
|
"toggle_url": "http://localhost/?_facet_date=created&created__date=2019-01-15",
|
||||||
|
"selected": False,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"value": "2019-01-17",
|
||||||
|
"label": "2019-01-17",
|
||||||
|
"count": 4,
|
||||||
|
"toggle_url": "http://localhost/?_facet_date=created&created__date=2019-01-17",
|
||||||
|
"selected": False,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"value": "2019-01-16",
|
||||||
|
"label": "2019-01-16",
|
||||||
|
"count": 3,
|
||||||
|
"toggle_url": "http://localhost/?_facet_date=created&created__date=2019-01-16",
|
||||||
|
"selected": False,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
"hideable": True,
|
||||||
|
"toggle_url": "/",
|
||||||
|
"truncated": False,
|
||||||
|
}
|
||||||
|
} == buckets
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue