2 rows where issue = 268453968

View and edit SQL

Suggested facets: created_at (date), updated_at (date)

id ▼ html_url issue_url node_id user created_at updated_at author_association body reactions issue performed_via_github_app
339382054 https://github.com/simonw/datasette/issues/37#issuecomment-339382054 https://api.github.com/repos/simonw/datasette/issues/37 MDEyOklzc3VlQ29tbWVudDMzOTM4MjA1NA== simonw 9599 2017-10-25T16:05:56Z 2017-10-25T16:05:56Z OWNER Could this be as simple as using the iterative JSON encoder and adding a yield statement in between each chunk? {"total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0} Ability to serialize massive JSON without blocking event loop 268453968  
636360861 https://github.com/simonw/datasette/issues/37#issuecomment-636360861 https://api.github.com/repos/simonw/datasette/issues/37 MDEyOklzc3VlQ29tbWVudDYzNjM2MDg2MQ== simonw 9599 2020-05-30T17:29:20Z 2020-05-30T17:29:20Z OWNER I'm not going to do this: 2.5 years later I have yet to run into anything that makes me think that JSON serialization performance is worth any extra work. {"total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0} Ability to serialize massive JSON without blocking event loop 268453968  

Advanced export

JSON shape: default, array, newline-delimited, object

CSV options:

CREATE TABLE [issue_comments] (
   [html_url] TEXT,
   [issue_url] TEXT,
   [id] INTEGER PRIMARY KEY,
   [node_id] TEXT,
   [user] INTEGER REFERENCES [users]([id]),
   [created_at] TEXT,
   [updated_at] TEXT,
   [author_association] TEXT,
   [body] TEXT,
   [reactions] TEXT,
   [issue] INTEGER REFERENCES [issues]([id])
, [performed_via_github_app] TEXT);
CREATE INDEX [idx_issue_comments_issue]
                ON [issue_comments] ([issue]);
CREATE INDEX [idx_issue_comments_user]
                ON [issue_comments] ([user]);