home / github

Menu
  • GraphQL API

issue_comments

Table actions
  • GraphQL API for issue_comments

4 rows where user = 208018

✎ View and edit SQL

This data as json, CSV (advanced)

Suggested facets: issue_url, issue, created_at (date), updated_at (date)

id ▼ html_url issue_url node_id user created_at updated_at author_association body reactions issue performed_via_github_app
1043609198 https://github.com/simonw/datasette/issues/327#issuecomment-1043609198 https://api.github.com/repos/simonw/datasette/issues/327 IC_kwDOBm6k_c4-NDZu dholth 208018 2022-02-17T23:21:36Z 2022-02-17T23:33:01Z NONE On fly.io. This particular database goes from 1.4GB to 200M. Slower, part of that might be having no `--inspect-file`? ``` $ datasette publish fly ... --generate-dir /tmp/deploy-this ... $ mksquashfs large.db large.squashfs $ rm large.db # don't accidentally put it in the image $ cat Dockerfile FROM python:3.8 COPY . /app WORKDIR /app ENV DATASETTE_SECRET 'xyzzy' RUN pip install -U datasette # RUN datasette inspect large.db --inspect-file inspect-data.json ENV PORT 8080 EXPOSE 8080 CMD mount -o loop -t squashfs large.squashfs /mnt; datasette serve --host 0.0.0.0 -i /mnt/large.db --cors --port $PORT ``` It would also be possible to copy the file onto the ~6GB available on the ephemeral container filesystem on startup. A little against the spirit of the thing? On this example the whole docker image is 2.42 GB and the squashfs version is 1.14 GB. {"total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0} Explore if SquashFS can be used to shrink size of packaged Docker containers 335200136  
1043626870 https://github.com/simonw/datasette/issues/327#issuecomment-1043626870 https://api.github.com/repos/simonw/datasette/issues/327 IC_kwDOBm6k_c4-NHt2 dholth 208018 2022-02-17T23:37:24Z 2022-02-17T23:37:24Z NONE On second thought any kind of quick-to-decompress-on-startup could be helpful if we're paying for the container registry and deployment bandwidth but not ephemeral storage. {"total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0} Explore if SquashFS can be used to shrink size of packaged Docker containers 335200136  
1065334891 https://github.com/simonw/datasette/issues/1634#issuecomment-1065334891 https://api.github.com/repos/simonw/datasette/issues/1634 IC_kwDOBm6k_c4_f7hr dholth 208018 2022-03-11T17:38:08Z 2022-03-11T17:38:08Z NONE I noticed the image was large when using fly. Is it possible to use a -slim base? {"total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0} Update Dockerfile generated by `datasette publish` 1131295060  
1105464661 https://github.com/simonw/datasette/pull/1574#issuecomment-1105464661 https://api.github.com/repos/simonw/datasette/issues/1574 IC_kwDOBm6k_c5B5A1V dholth 208018 2022-04-21T16:51:24Z 2022-04-21T16:51:24Z NONE tfw you have more ephemeral storage than upstream bandwidth ``` FROM python:3.10-slim AS base RUN apt update && apt -y install zstd ENV DATASETTE_SECRET 'sosecret' RUN --mount=type=cache,target=/root/.cache/pip pip install -U datasette datasette-pretty-json datasette-graphql ENV PORT 8080 EXPOSE 8080 FROM base AS pack COPY . /app WORKDIR /app RUN datasette inspect --inspect-file inspect-data.json RUN zstd --rm *.db FROM base AS unpack COPY --from=pack /app /app WORKDIR /app CMD ["/bin/bash", "-c", "shopt -s nullglob && zstd --rm -d *.db.zst && datasette serve --host 0.0.0.0 --cors --inspect-file inspect-data.json --metadata metadata.json --create --port $PORT *.db"] ``` {"total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0} introduce new option for datasette package to use a slim base image 1084193403  

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
, [performed_via_github_app] TEXT);
CREATE INDEX [idx_issue_comments_issue]
                ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
                ON [issue_comments] ([user]);
Powered by Datasette · Queries took 40.953ms · About: simonw/datasette-graphql